var/home/core/zuul-output/0000755000175000017500000000000015157205401014525 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015157214702015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000232774315157214527020301 0ustar corecoreWikubelet.lognc9r~EYd` \-Hږ%C{sg5݁ϑ(Ӄis$WU)X /{b}Wߟ/v׻%oysシo_fZsZ-jC4%_̾zׇϘէoW7_~uyi{|||F^lWo%vz_/-~I|g\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? r.|I"n j/\U R[EC 7g/7_E'c/z&BBb}1lχtO Ң` w"==:zR9P p&=;mk b2 ܨ;RJñF%QV5pDVHwԡ/.2h{ qۀK8yed3KdMvw×`21ɻ}/<@8(PN_,_0;_x+Fyh\dN9:bġ7 -,qӸZpə+1i:yWO[l6ro%-9}tX 1hɨ|аK"B[T'A25T jzdz6"ٍ߬i3x-QI)k=-\$)'/N̔ڧO1_9C'/-#\1, Gꨦ#Jҡ/츣s6KT[<~6շ}~90}(*T7siv`̞Qj_P\ Q]GcPN:E↖t8m1ga5"f[B[fhTQkwW/+b8_b|r wFuRI%TѬza\_3/wn>- MRyC jnz+(%2݇^^J"ͦ>CMMQQ؏*ΧL ߞNPi?$;g&uw8~Y >hl%}Р`sMC@ztԝp ,}Nptt%q6& ND lM?ָPZGa(X(2*91n,5o0/-y+,RO }S)G,K>:R€A`*t) O ]/* @(*<[ǦX iJ D mD~q2볯Qfq&}͕C,RizpV:!җӣ D>P.BvJ>mIyVVTF% tFL-*$tZm2AČAE9ϯ~ihFf&6"W&\jVJ5&jNgB|90v߁R:~U jڞU~oN9菏xԞJ|d`V)F5d,0SSNK9ް4:ÒozsB<^/鄌4:B%cXhK I}!5 YM%<>"۞)Za@Ι}YJz{ɛrm4^jC d-saܺCY "D^&M){ߙ>:i V4nQi1h$Zb)ŠȃAݢCr|<~gQwQ!q/C>*({bʂ!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8?,0-Y8Uav0NET݃jYAT` &AD]Ax95mvXYs"(A+/+o+{b]}@UP*5ì"Ml؈W|sN{mL=@N'DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ {f'}KXg%q3Ifq CXReQP2$TbgK ؾ#AZ9 K>UHkZ;oﴍ8M*a~ff~6|Y,d,`!qIv꜒"T[1!I!NwLv}\|s.|=o4_P\(Lۻ5_vJBR_v'5n[lGhNU˿ۜ6C9OC7Osnkje*;iΓF[^>؃n糨I@[ tWv Few9J֥WmN^<.eܢMρ'JÖŢո%gQ=p2YaI"&ư%# y}ùXz!bm5uAߙXc90ov,{*9ߎ% qƦat:F=uNvD=mdZB4']a.QO:#'6RE'E3 */HAYk|z|ؾPQgOJӚ:ƞŵ׉5'Hdg;md^6%rd9#_~2:Y`&US tDkQ;>" ؾ:?/GSZ;m#Nvj{,4xPA*1bv^JLj&DY3#-1*I+g8a@(*%kX{ Zۻ#es=oi_)qb@esQT*:ÊxtŨ!u}$K6tem@t):êtx: `)L`m GƂ%k1羨(zv:U!2`cV, lNdV5m$/KFS#0gLNO6¨h=z,ffK2Zu8C1}PcIr.e'.I(NJKڨzHJodo\4"+(Nck!a}P `x ăil_=!X;__ZNۈO;j -ݙrbH9t֥s`Mh,욕v`;VI&ϊх{211 VSxSew{~n}Lr7+Tcj.R2>e3V#b,PY0TEu1L/]LTB4$`H6NI\b*AYA\(u|@~*8n]rMQ ;N>Sr??Ӽ_/khSQƷL@4;嶑, }t&&\5u17\I@ 5O? ʴ(aPqPϟ'VO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*%WyΈ W\Of[=߰+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^sнM\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bd+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTtx袥:2JޚݶKd7UZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOa9C1L ={fm&'^tigk$DA' elW@Tiv{ !]oBLKJO*t*\n-iȚ4`{x_z;j3Xh ׄ?xt.o:`x^d~0u$ v48 0_ | E"Hd"H`A0&dY3 ً[fctWF_hdxMUY.b=eaI3Z=᢬-'~DWc;j FRrI5%N/K;Dk rCbm7чsSW_8g{RY.~XfEߪg:smBi1 YBX4),[c^54Sg(s$sN' 88`wC3TE+A\.ԍל9 y{͝BxG&JS meT;{З>'[LR"w F05N<&AJ3DA0ʄ4(zTUWDdE3̻l^-Xw3Fɀ{B-~.h+U8 i1b8wؖ#~zQ`/L 9#Pu/<4A L<KL U(Ee'sCcq !Ȥ4΍ +aM(VldX ][T !Ȱ|HN~6y,⒊)$e{)SR#kהyϛ7^i58f4PmB8 Y{qeφvk73:1@ƛ.{f8IGv*1藺yx27M=>+VnG;\</~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނL8]iSCQ&s~In/SZ % 'I Ƿ$M6rN+LxE>^DݮEڬTk1+trǴ5RHİ{qJ\}X` >+%ni3+(0m8HЭ*zAep!*)jxG:Up~gfu#x~ .2ןGRLIۘT==!TlN3ӆv%#oV}N~ˊc,_,=COU C],Ϣa!L}sy}u\0U'&2ihbvz=.ӟk ez\ƚO; -%M>AzzGvݑT58ry\wW|~3Ԟ_f&OC"msht: rF<SYi&It1!ʐDN q$0Y&Hv]9Zq=N1/u&%].]y#z18m@n1YHR=53hHT( Q(e@-#!'^AK$wTg1!H$|HBTf̋ Y@Mwq[Fī h[W,Ê=j8&d ԋU.I{7O=%iG|xqBչ̋@1+^.r%V12, _&/j"2@+ wm 4\xNtˆ;1ditQyc,m+-!sFɸv'IJ-tH{ "KFnLRH+H6Er$igsϦ>QKwҰ]Mfj8dqV+"/fC Q`B 6כy^SL[bJgW^;zA6hrH#< 1= F8) 򃟤,ŏd7>WKĉ~b2KQdk6՛tgYͼ#$eooԦ=#&d.09DHN>AK|s:.HDŽ">#%zNEt"tLvfkB|rN`)81 &ӭsēj\4iO,H̎<ߥ諵z/f]v2 0t[U;;+8&b=zwɓJ``FiQg9XʐoHKFϗ;gQZg܉?^_ XC.l.;oX]}:>3K0R|WD\hnZm֏op};ԫ^(fL}0/E>ƥN7OQ.8[ʔh,Rt:p<0-ʁקiߟt[A3)i>3Z i򩸉*ΏlA" &:1;O]-wgϊ)hn&i'v"/ͤqr@8!̴G~7u5/>HB)iYBAXKL =Z@ >lN%hwiiUsIA8Y&=*2 5I bHb3Lh!ޒh7YJt*CyJÄFKKùMt}.l^]El>NK|//f&!B {&g\,}F)L b߀My6Õw7[{Gqzfz3_X !xJ8T<2!)^_ďǂ.\-d)Kl1헐Z1WMʜ5$)M1Lʳsw5ǫR^v|t$VȖA+Lܑ,Җ#+sM/ѭy)_ÕNvc*@k]ן;trȫpeoxӻo_nfz6ؘҊ?b*bj^Tc?m%3-$h`EbDC;.j0X1dR? ^}Ծե4NI ܓR{Omu/~+^K9>lIxpI"wS S 'MV+Z:H2d,P4J8 L72?og1>b$]ObsKx̊y`bE&>XYs䀚EƂ@K?n>lhTm' nܡvO+0fqf٠r,$/Zt-1-dė}2Or@3?]^ʧM <mBɃkQ }^an.Fg86}I h5&XӘ8,>b _ z>9!Z>gUŞ}xTL̵ F8ՅX/!gqwߑZȖF 3U>gCCY Hsc`% s8,A_R$קQM17h\EL#w@>omJ/ŵ_iݼGw eIJipFrO{uqy/]c 2ėi_e}L~5&lҬt񗽐0/λL[H* JzeMlTr &|R 2ӗh$cdk?vy̦7]Ạ8ph?z]W_MqKJ> QA^"nYG0_8`N 7{Puٽ/}3ymGqF8RŔ.MMWrO»HzC7ݴLLƓxxi2mW4*@`tF)Ċ+@@t޹na4p9/B@Dvܫs;/f֚Znϻ-MHVuV_K2k*`cKxuBG&24T}Lai 0Va(7K#ӊ!,ZDxFQO*lם>!4ӥ2 ]8â6 U`V%`!c%؎ʨTzrKh! c.}.D>)d_ 8rcu,wf2?Ǡ*_lDn}rauyFp*ɨ:UiM2r:9ct X1lmĪ o玓,R%!`hGT LYF#g<cm${|Xdu4tmtїUJ\~dc0KcMlf2?mμQ ߉J4WrSHTdp"ӹ'cJq2zPlX̯.0H!ND@UapVoGڧD5>H]f@!=߸2V%Z 0"G4ȇʩ@]>Y$ًF_Mm_Tt)ib+q&EXFu򾬳ǝ/RS>r,C2NfOjpcm{Ll9vQOT>9U;])>6JdbXԠ `Z#_+D[7IIjJɟUh ҙ"`"a ߒ"G̾H`6yiCk(OA/$ ^%K^+(Vr[RR1"u4A.1X0=7f/"(o9/L1X{]q`Ȝ/; 9a>E)XOS K9mUxBa"'4T[Jl /K/9,rlCAj_TiǘP,:4F%_0E5IE'rX-|_W8ʐ/=ӹjhO%>| :S Px„*3_y.g9| ;b`w NtZtc> ײ1KĴ{3Gl& KT1ZWX8?C]~We$9; -.D087?1a@P5B,c}jcGȱ WW/ @a#LA4.ٹ^XڋXٝ:^Izq. ٽƎDn6ٹBc5Lt;3#i3RAٽ9| cbpcTfp> 6L/_x 'ۙz7~w~);qU9GDT! 6]c_:VlnEUdn6UˇKU;V`JUݵޙEO[)ܶCy*8¢/[cչjx&? ՃJȚ9!j[~[' "ssTV2i sLq>z@JM->=@NỲ\쀜*/) ̞r21.y? bO]3?C!yw3ޯL_Su>o>&lrw&i"< :]_<<7U_~z5є/rfn͝MLmc 6&)e+n7cyy{_~궼07R7wPuqpqo{ߟ+[w_uOq?u-|?WS_tOq?Eu-L_p?Cz .e ϿO*3 `Ђ6a-`kIf-s,RL-R`1eL~dپ&+IhYRczr?㐟,v~,b6)up)3K,RLW"Qd9JgT\1f3@Kh% a4x,kA k ^d kYj5Ah𚄓vXZhX1xҖ51Y +Id ZZ\C| fD>hB֡#-$+Jpሟ,Cg:6 3 xH "}C[`ӨOAFn5ʬLHϰ:N@VcyBI#Dr. "h hg ۃm-qu>V&൘ G7qi#^tҒ[JI!{q*lrD܇Gk@;oI<5xZ4xM"؇'k!>V|lk'{d+ :sXӄc)?W`*|\v aVT0"tMًcΒVz]T.C$cEp._0M`AlF̤@U' u,—rw=3}resLV&ԙy=Ejl1#XX۾;R;+[$4pjfљ lݍ3)`xvcZRT\%fNV Q)nsX }plMa~;Wi+f{v%Ζ/K 8WPll{f_WJ|8(A ä>nl"jF;/-R9~ {^'##AA:s`uih F% [U۴"qkjXS~+(f?TT)*qy+QR"tJ8۷)'3J1>pnVGITq3J&J0CQ v&P_񾅶X/)T/ϧ+GJzApU]<:Yn\~%&58IS)`0効<9ViCbw!bX%E+o*ƾtNU*v-zߞϢ +4 {e6J697@28MZXc Ub+A_Aܲ'SoO1ۀS`*f'r[8ݝYvjҩJ;}]|Bޙǖߔ 3\ a-`slԵ怕e7ːزoW|A\Qu&'9~ l|`pΕ [Q =r#vQu0 M.1%]vRat'IIc(Irw~Z"+A<sX4*X FVGA<^^7 vq&EwQű:؁6y\QbR9GuB/S5^fa;N(hz)}_vq@nu@$_DVH|08W12e_ʿd{xlzUܝlNDU j>zƖݗ&!jC`@ qэ-V Rt2m%K6dX)"]lj齔{oY:8VmS!:Wh#O0} :OVGL.xllT_oqqqLec2p;Ndck[ Rh6T#0H Q}ppS@ώ@#gƖ8sѹ e^ CZLu+."T#yrHhlكʼE-X'I^=bKߙԘ1"+< gb`[c1髰?(o$[eR6uOœ-m~)-&>883\6y 8V -qrG]~.3jsqY~ sjZ+9[rAJsT=~#02ݬf¸9Xe>sY~ ae9} x* zjC.5Wg󵸊y!1U:pU!ƔCm-7^w]斻~[hW$k sE0ڊSq:+EKٕ|dvvjjy6 æ/ML-yz,ZlQ^oAn-})xǺǍ--qcl:WLg ӁvJ[ǧc~Of+8qpçco#rCtKӫce0!Y-+cxMK-H_2:Uu*corD~@N`#m~R:ߙ歼!IZ5>H;0ޤ:\Tq]_\_>e˲\oUQ\Wߋ47WwߋKpwSSۘF,nC.\UߋoVEuY]^VW0R=<ު˜˻ x}[ێ'|;c^ M7 >5\-> m-8NJ\ALd!>_:h/NAC;?_ξqĎ6xMY(=ͯl~l8V0٨T zL{Ac:&$ ^CpH*DW\r2aR|=(L X1|wrO_g ux1^^V2޲jMi^b``Q#dBxV#NBk1;DAV$"*1]Y~ d->'I`ػ޶%WK(~ ``g&c83`AnJHٞ_U$%ˎULXD.ֵKFܨ(I@ U<_DUa2,jSW%4_3jP0>.< }UR U,RU}S RT@%d{k=!UqSVssk*~v%_~}WC[5Ps9H`ﭥxP8v[ U=xy]FWUn,QNbPd"ݻ d__U O=L k8x/ǰE5{yB2!9`j?!wNr_%2SRmlNI7n8:Bx]úl(vZ B5qPI] K[G;9~ԏL׊j&p֝%K`k# \!=3F 4N֗yyZiW@ a^+6/RDۋT]Iq|Wub ȩd(TjG;K_;A>AOZUeS+Ob8l~2ӇƧzڞ,M&YÊħj!yWۜW V|l4Ϛ>8UH\b^bMih}Z!a}jA~J[oX}W8b]SL]1yg-Q`<ӥC$&j[և&p{M:j|j_4 b7< ϶j+4|+w;qyo`Tz]+({fcϊ^0GB(KQ0b[h xͦz(ƪe5] {{)Ty*nqwѹMkz!cuS`^w.$X\~b  6% >]^4PۯĠ9d*!!{[]wMйLO;\8Lj$LO q8a2vZ>CP^40םFW}9W$) i.B owyw(QqǴ͖@yT轪};I0=9~ˏ!cR\2Hb ) D0Wxs|F+~UJFBc{grS^D(ULK&M/ME1`QfD4͵g*e Ao޼; Q[2稜FU,) `jПrc3s;S:BZ - M./z)1jR}hO!'W۷" 5TFҴ>4L"㵻@g%i2 teZ՞hhMFT,U6`\-խ@W/Qqh0`("<+JP\#ч_vRь\5$煱pǦp[\w5c6uύ"ח͢$a&"rcq$2#ۊyDv| `"4$Ln>&h ^ ApJ\KDЁެ$SPcmIzY7 TCdiʹZ tK^ Um u&:dYzqh:mpSPi}E3)>Dk;]WwyN9c9R禖LnH+҈A7ӢVC]gE"-<]mѽCvV[`1ãK]EbC]qvU6Yq9 HKݫOa, Rom4 K mTțz0mV$XC=մ@ /9}2\4i:D.41[֌opF6@ć(¸7?_`it3(XU]  ,t(^K5t>~7D=; {tI{Z-V֌1gE=Ƥ3UU#St)^q#q;3͡mɹ,og}־a9`Mi՝O-179UȈvz2MWO+H2 YtzMRyze[[XI9n9Z%in- ΤQ5VZKv@Kc 5y>xX6~dsoi|g#lma_ڱl29!t:"9%☻2xO4pKPxcP'Rn\PŸe%L(w O?5eٻ4BEx~@_26[o k57~w]7 Xlg hF"ˆSu%_|(ѓcI#C,IB9:"4VDB Ǒ@S57go)mwበ>3v@Um$C>9Ww(\XGwdzE`{YÏVN/wvaL-v`9 }Yr`!Lo`1ve(,mdbWNqH ;Fa!??{۶ϔ\+\'i։?;ih")DHɎ/~3$zؑtM7yh*$ϭ !nG>e.Ϗp%/1y ʧѧwyte"py(fu|Ipj̬0:ůF$.SwOt?tߚeVU}u 'sy/3iݻhH 4'QU7Tq(",Q6x qe7tЁNt+P‡T0鍟 8&\c1+C?O˟M0{g]/=xn>m56]Cn] of`ş~@sxsbq*xG!`^t"N{XHF@{\HܞK~f3#(%G( @0>ő O^b ؠO |uǕ0J_7Cǣ I B=IX4s" y=uՏ S9z4=wϦ ru4Xu~i>OGݾ0<fiĕw޳ە=fxVXoh.=_W(8=ЉpA:ΰFq<~s O@6KOסӰXv%iX-4OrOi--/^ʾB|+ݧN,#>1ӷ{gno} Wc| _Nc|׈d{`-_g'no|} [eN4t{5OgߏdY0q]]Zs/yO:n&:[A߲(-bӞ$6M,:Yv{q NЋU\k`O5bI;Ӄhp5eFqtO=6Hi (ct7:Cӿ:^J-Cs8d0R,] *R´:v&{}ŵ,Vkf˲,Jeyh̢rN1_$Ub.Œx)xxI6Xg,$24@C L6akmk'y * -2 :LP ʂ JiXyC&h$.JԬȀ)&7)^  6z0a̢$㲘cchfgS ӫ,juiUX ´E)+WM;j*ri{{'Ey"C=JMYɬ1vFv+y❁si-oȷGoW&X}{ e/QU?C>Q@+}Xmٙy܃umtpѪ2[dk7nSy@o iQnƨũo=FdۻWB@7l RlM{P,kv(&ݤ X}s:[iBI,ޑ.n**03  >ŬIҮ+j'k\N^9_-xؾW 9%7$2h+򀻭k@}]GRػ<h[P}*:PQS`p?e%Jw N(}tOB. )eLJ9ta;4 cvetUjw~;@;u$فPJ(ہP;q= e;@;u$݁PoPoB G';N8B= w w N(|OB;Nh8B= K:Bn4%Ǽ͉`buwxԹ o/rf㩸MV%VGKFnu@*nF3-+)AY2XJz&:F/$EѤP6Lf؃t\']z 3)̣,O7UShN^"2R[/>̛ ( kԂ]"/P }rIn<8O茪(uhvgaShRb,!Hm DK?q*M t1tк4_E`A{U;F͐ i{USjLVZ݇0h>F;1ԖgRJ9EPEAJ|7x7BpFzXTկ`PM80uQA,G$ כ Cܵ5s[JQcClâ/2|'e@[΁yrHbc\w ܙ^H>y^sˁ2e8>~1+QjF/&B<[@_eMtvN25R`Nfn*.by +.u1lt+(,k Z- 5a}^G9<=,ZD"c%&1,Xe0+84DT@A;@2m}$L*Sm/a YHOy'3 EсCuTTlglDRe Õ\=e:-}R{W0up,2IW+$cz 3s a"$?s0:܏p$s8ydg8/uCYG&5 4iu:iO^Guv 'g)7ӗ< +<.a~wVjY-T"Jo)ePIjP 9fG,~&{L_paY5u%|= GQrxI2ɨV_ ۯb;f)G7"~mԟEj$n)kԊDH2z;DE57[R|6|\Q`4.΂w'*Ӄcw8ǒ6CP9zXXS^7O$ 6+ 4g$i P7`T Ҽ.[M\%y6I¯-FRV6-~ ,rp͕ޘ5D ۊ4ODlA!Zؚ=eeUȁѢdF ]*J*[F6Շ`a lW omaNx'"2UΈHі}">"1*_ v(SK)b1=xG4R/N%|j}5 ju%m`CSfj+@s#¾}GIW]6m +#^]6N]HKdmR\(r-[ar)-Nl1AkIiPۡWnFU,p神%C՛hr(˴V:8U:>?1t P. =;M8T84NdK؏-p8H%cUwݿףߎ}K}'1Lp~QBj7o /}78Bs{tɿ7p]oG=vr֞y[=wr8x2s{~8+9Ś?Cۃ]詅w7''.LI9=UrTlLN$ja٧2)a:W$vfchYsO'n|X=GkxA+xLkYZ0Ѻ*R_1/Q`hK{<.YN2Q8N1tQN$8N<8&!f~kri̖(}h%eC]|} D4ZF"&Ua>rRLHYAG~ň9xYXX8Bg^/o*‘#_$8?GQM3R Tנ G+SB%x8ˌpNp+fPD_4: ,36愶I:΍:% M癍`J[pj=7#$8c$];7/G=ZF/Yӥd[hLdbd#4{;ehN'u`B.Kw1XĄCLcqj&1Qu.Vz^ # Èauwߏ,X v<1Nݴ_ut*r= n_-{GY/a5t?fY"Wg[.#R41ęU΍mPܧkd ,$(R8LV#t@x axm0㔀!(AV 8åi4BuiI͎ɸ EQy=Y\I,S2c QVrpIyL(<Г,\P@sq} $AgҢ> ͒p*-g1Y3~N?eiOݽԸ>>%/G"؄Ym0g;p[>B$B@0,9i&a9v\ha{2Q \kbL|Am0M E(W7DF":,X5H"G!fDKͶPv%O.B&y/Fg5U2H퇑ĎKHCfL&- Yw8$n[t7s'l7cƺsf8tP5PO0(,DkɑWAk T\: ;{1=ƯٰǤL"Rx%ΡբN-B 0VlY~[?~O7Mx w6,jBmL}A׻{ `k6gЍi!=l2WZ+YK1ݲfQfbTAv`$])vFkQxуe@bOHct0(ppń$e}BS3OR!o@+rؐhw9{K+r5,Z+TPY0Ds $_Tiq1[e^FÔ"gks |I}c{w^8wqP3}E"d-1I$pr( +pCJZVXBҕ?8anp2^\H(WV߈I =HXytOׅ7RK .1QB +2t&Id9iڔ3w߃0tHZV}Z2Q|x<0K Nq*1al-#: !q_XѪ<% (rO5MH"\J/+_8% j387?YͅI_'k-_ĄACqU7F8l p_#R* *x78F{WQrrf4g nIr j ܍'į{$KC\L:<`gzIP'wMxl,&$pZ. IcQ |NGkn͈<Ӄ]~C{$8^3"'teL+^XpF Ah,Xu" I<7)r!}"qoym v! ;~)p4\l~LH1#o*dL˞=ɱR.U+{vGEN`D`*II &)}mK^I[N[5Vi,L)@SĖJƙR*B,IBl%CIW9Y)&▷se;ع[tENyx&)N6* Q<6_bu̕= $Xeg,16h1;a"ۣfЎN:oM_Zy$נdHї.m%<2_THr 5nalL9pF.q!v兤GY}4s+rHp\"eߴ[ޔNԪ&sӸ[T2LYg2Y^`y."v6$DW΂!EҠ5N;ާ2:LF ۯ6]w[$użޕ̶l;$HOZ0sW- w5Q^s 1߯6o8KY17λQq^o! v1 jYt.n6*aL?H0G3wCJL$sťcG[Qܧkdx$5wSQ"$ƫtùgD1;ѯ:eZ48zs`z7~U[ex(Wf4 ̵WVBNI6ː jp͸W%>Xӛtd*T U3c1hhB[xѾ1ӊeL9.yc$]Y+48&Л9cț]bh|T3&Nf/5Ml1a@8xzpPFjε#Z+I7;3f&%Lᣳbs.Gm||Y \'J)ez ~]ۿ&qd[ c(=}|ܮ랺ZgZ$ŅÉqi]AE \m <  +@lhyfθ2msTSF y,PXG?ò}ޭ^%ZԶ f Y݈nvnز!#wϵ |Щwk^j3KcZ 7x$ZFHb_ׇ{Vb}ЎF9Fr;y(y#Ȣ3/P紺!8!F7+VػeỈRށ],'V(Y+H:"χ5zq`X|,fiUV4 KPs't.ꈚ1hd'+o0Ys&=|di-y6Թ-enDڊ⧘R笠>nXݺ< d_{҆fnь^9vUc* ΠaAK&@A=#~gRW77XıL'5 '$}3Bs$=\]}=5dqH@%#yլOg3` ty\jIx|H÷Fn{'dC~ҽ>ܓFS1 / ^3?X3 ,Xcqk=F9pn,D#Kě c1b VڪfPwtCzz҃2+&@'bIxo.7WCy/d vN4aâ'|)뤍S VxWiGDlhjgL&u &"t)Z;dz˜eQ O JG[ΐ 4ƚl|4W. Хv<2i ]hj "$Q~YMrs{U)h[7]E勃l!Q mwo;-uACfQ#%1UnQ6gд6 \(b\tUS&ck%X'C :%&5ö1?tucw;zU F$ndR310)KF_GV/4g:E5}j'{8j2?FGy/{Wȍ&>xld180${]<EvjHrK6Sj_"IK8E}[]3+]Bߓ/xvLRy>|EIzqv`qcpӾv3*+)RČ^;~}TLZ ql0ϓ*.9iWe, v F׵&v%N>1zuiqH ;\T!<kAWg?OP n^׿_53?(=mt{Bߧθc@{ўk砞@np Kg84' <utOZgWDMʍK$jI$+[chJZq7//N>vƗIy n-Qn+@P)߂G; ͭUaɹ=CѠ)IL@%#):=?^zz |3N.͕ڏ|«7u; >y$Bdk쯋7 +BVl-NCSqq83 N2\toW]q& {r<AxA,TF~  ЎϿvz G+!_@ ډq0#r"6^wxG':ț Υq1| iܿ4!tɛh0*d lyG3wۆ-hNfph=D4]ыFc|[|DN% UO,Q nܼgyՓĨxva4`*qDAClNFp= zќtėCvފ SZoYY(;YwTglSlqIxε@#jgYqt/Ff/LbLj[Wsrra ֳ: 9yS3|xeڷnpKHgܸ Y&A!NmM8I  A\f$Q"Z\c24˓w;qixEt岖p2{R W|'` ȻuI7&_fW7!0pj.&EvBPE'i^ϔ1EӮ3gm Fd׭2@NBO~k e'^z7CnP㲀kMpFܭBⱲws{ճ=^ x~ ԕS,,Ucf΂J7e fDӅb bҤ?#na~H#%o}AŌyq];Q +*i, b u#d #k150v[v8k@x) VQTAj PU6j BYM0趓9:b")Q<@F Z]k{L&1&8YKJUV6nshPrHe{-5YMj}So¨B Q &MJ@\$ka֠Da内: x X\$]or?;8?H>vdC)KQN])@bh 7ǟ6Z&ztOZ ^B$.׃[brx󧷓PuW7 .*'bjQִ4<7|J$>ՆwPtLuD`-N 9c7M [_ZC.k=4ZЎP3G݁5p!jG=QgVH-Sm08exPÅvN@Ut;u,@[.c!c) +ý:,!(R{Ծ6ۙ;M((6٬*m6UA^OKig2<<̽a35P+m ;9Z85`Wn& :Q|5g .M.xQAK|Gz8LᄨJD8emp1qי \pqynPބ_?]}%1 …n('hFU?zZfh*00f Ab3ʔk#68c}̹s8&8:}7Ddݎj\as H.q8H^WM2EzUj 3S)|kf(ʥ· $tiܢTZT-nG,n<E+۲q߼hyy&V@F2a]f30g̥`mzlV線㷚װbh~⋶XUܵ/:tH: njQ&$؁K҄ l)fL2ҕmnGln Mxw΂K*a A SF*O0;-icR `b$j2̍]nu_^ZQb}fUof%8]'a& cҜI .k̡KAZ_(bOvokW.Rـ;=) v ~w`6=?ʏ!'+# NX߮Qvz=\:BQ9 >UO8B_M~T !r7hMbxPZΜd9" }=>@x2@jA)xEGQgG޹v':q])VJcdǕq|}(d8IˀƝ(n%"1آd_)7:!ao]2}rq>.oZ!K+_M#E ީ>׀G?W\6׽5gk)f6W\65 u'Sk) 4k9-%*hk6JSP-ZB@T:ݤW]kyx+xTcC;=JxW /#Ս+w%Ϸ C`q%7 5A~uƾ${溘JLʼ恩l*LQ *N~!h `B!L ;`!َ_Q8h^%wSo?_NWwE;[ߝkX6v*oKi?CsfƗț 7 PNFP?>.:(0J qaiCq ;;>0? 4*_Af 4`ż;QmwN]N&ن7`ܼwA{2{ێincםȋO-#۝٣ӳ 9[0[]?zwr5 iQ9:乯#?CJʭqȫG+`D.w)G`~ek8ŎVH&v2 z[ĥ;>y=i2: Iz-6

ZI6njqCa!o}Hhla9gUEec/#W|րLR)>qٱq\@/Aϊ²N0D+E˅f JEžkOSR̰R$m2Ab…!ܒ0SaCN2~h չ}*ԎdD7yMuP[k@vy츞8v\O\Fv;#-c_e^ 4eRF!%YY*hjc!c󆮎SJmC.MdZ ,O  S0 -lNg$1GtD 蔸Lh< ~rG4sʥYSAL $RO!RGR*mʣzJ,4Rm<aeP1u*Mi&SfR$UL!M6k i%mqnӍ>:LNS3g&7|*^WCn1h2{r4W f2wy*4 .UayiNK&\nAm 22(Jȧ!wkEyɻ"AFJ#bsOa* Fvuim+i/vj1LO mw klt4s]VY1c R1)kJw;֠rfVT.ǟE5=*ױuOП,P*>m|(n_͙QeʙE3a-\"e<ɸ-]!3aˤtp$>mN\D0zhVs)RJRTNwGTZ1 ERhBV搗9P -e u (d)*\wDU) [j<jQRHL52 N$4+QJpxYN-E4\NX :` ޗ Gi_E#jTt8ױݭMOg^]/Fٗ ӍLl5[Rs}oooozq'\C=iM/=骸9NyYqpyq75נ#+|d1Lͫq*_^o#B; h+`,޻G?ϙvAɏPob㿳v3w8SQ۵zO:P7e8J;)t&<6Lfi@0L2Jgi`d ֘X4Yz!ߵ`MH)(zJuhep( y!#|?)r} jг"X'[$BZw46/\"TZ ݲkX:kBka"߮i(6Z[k L+Ylg柇k#*h \Xܜ$RtE,HxwRNXNYkp ři(RCT? t1[NeDɈC=i 0U`:y0.c$nY%kp65p3Dqa$`cxBuh5N->ͫ`}7ŝ;6_2|4q{aֹx{9?aV_@jS=šC`;'Lfƒe`}E  jm vKqh;'|/9zܮh` /0©F!\,̣Gfypbn|i)?GzѴ_pQoL(yЯx\IJ-񞩡C.izV9DɡaUR/'#_OL`ZC^5'6)aei0K9% JEųyA㯪?3޿?g_&?ߎ=F=NVă"B_6a|1AqKX.MԎMT7JS€3l5cό/Γt)r2|Zp/+uV 'CfO Ĵ\x4^?{NSB0/`/߿cΪJ2qoh艸}GG3ϋ~<:~v|T?ٹC ǓtpѣNk&W/[ktz&.Q?]Lk~,GC˷mt1[ٳ5n:[ KN lTφ S0H !W'.=1+ڢ)\Fg }vPB[Yg UИj s}-4k;'?'6hY.iR(rniY.Qu[d%q1xP@S86+ACkn4ն6MBktO~ü.9Vmfg˨ѵabJ8܊W۸(-k0d Yլ h/ۻV5G&/i8y@\.$RAtl"`'#Gc\#fJ)юhlqJGDpSƥ4[cm}<R$&TB &03p>h3}-km"@Fު뗎T4 VH`݇+cd0ӧOo0  rJG+|$ HT@hVB7!P!ި>ćh!e:͍5 H]H0ѹ)5c2mm"\ -ofB1ס>[AhqѰĀ,hQSP(U Gyvp:gh OATEш |홥[8e{fVju f>T=}4>Gن3tSqꌧ1t/pmbUيpwkGFv;P@6*zJԣHsS굑RF|'<L3>`Ja7YBBZ'ozU479[j[ eM4﬙ё~eѫq=W.-sVNSY\!Im,(̖E,/Tc`Gvpg}Gu7x<ϲi3dnhpo1IuKxPu zu ~`1=؆t] Fj@V5v_@Ni˲>8QswbkVd;Up? > Q7E$` qxxo>X>gL;3f( Ǜ三x4ui:ϳLx!PHc+y,$Z FB<<LJKl(B6[ TskOi'AE|xSM078mo N! $҂#~-@T•+yJe `$)462) nlK|OmFGWªH HkVq+~+PIq_cJ2v#Z`n t/_>_!( ΖaJY.K/֚g<\p+k,{aBs9~̉k' I:R`T3`"_K"…24FVPߋT9 ߜ>ueYN)D;g+i=qXfEp E}ewhű%iW~:2hv;bOd9;)U䎥RNEQhWHg:@0h[0h0r[,Wo&mF7Ŏd~E;B4hY$ȋ,؋, Z!M=Fie%}K/& Rj45'JB.pd]7U*YUƽ6UJҪ kaA5QV}hշq-El)n]|O|>Uma.wo]Rs6ԂVϻ̀ , jϗ:d94 5Z~0yq5|_70_}ߢM??{Wu2Ə/}.Ck1w=VH=eYW/SZֵa HmI0χDlm\MmZ'n6 l^nBIZMBP* `%%n/A8yp<}@^M>^..뮬fE}ZɣG#\6FX2ՙh*Օhhc=pz*6@ ;5" e¯i~ kjs)Qb\^bsJ# d$~LnoIm=,N,ne;V5jxO]gF*oT>JMsB!ClE&q]&,( lF=A\\lRxƌkmN=Sl h3v?{ba(LKLrV,{$_LjaꍡNs߇>)iP'L3mlL )+)_ܔ,3%׮ާ4qIxN 9h%6S.uMC{I'G# r0 n_dL%+Z>8KnɊ|1?\ް7h9HxZmM1=ݡRyȽ$DR𐃙iv: +K5_TS9NUDŽ̧*Tζ_;`!Yb8n]p`wc#dg|-O4neD1@F]"^<3Z/K MP=:IzTFFsI"QVۻ@T{ ik_dj{G+R{֦:0wl1iQ!SHYiU%rաTA8ORJrK޽2x,2ՊWԋj{CK1jė/RH[{k{Sfd8oKWTj{G F3&a#S fi_j{ غQrTHH[{k{QP"%c{R*T痶f=rj:=&SGʟFekҦ {|먓ewﯻ%!u>ʇ il 5>P` {`w=gNdtOҀ?Ȁ~mr/iqt3p{^ iۙƊg1jcTLVab8p`#uQ<ӆNHw^u_M`PL|iv{.'7,i 0b8g)[7X]I{}_t"q'mړ-_\ӊL7>'ȯ.uN7To.8]Ǒ|lczu캋7ʠi-؃榋^ &S*WQiFp5\!-}2((pMqS/DKZphj3ڦ$nXa&e(PWeP(޷IX%N#X&Nzϙtty>SғD (wo`etv׸9Dێ%!7O"AFAov^# ˿9݌ Pj oQ%W2sLUX}؞-}4BQ$ͭ$1bT+7R&ʍ6nXo|rECx'i$V^;^s6Y`mZ7ڢQ[Y_w=Z2֧53T~2㔸Ig)FȒiSgFgR[&BSg>nj g8.ϻFw;Ș}ȉQ &X9k-Z1Y鬵iaq1VB9RGP."V{Z4 4,ȒdrHZ  xU h/dJpy~Y:`]@-WZ/Up?T^V}#4JDaM6Z28q29{.UF:VsZ]lLy̽)6zob$.4j89 Ȥt(7do)uy_'ۻ|*B㍫_ixXH$vYyh./d]t^kww2dw2BR& xk<<^ 6> V»л֏-Ǜ86k Yo)g+#H1S#8_W/[//!D@^ge:H9|aϵF6ŗF |]1ts`WCPT'E Ji+vMsƃ2im^G)*gSFrx<oTͮy}+Hf?%!C gTFbx+^_麞ǮҬ}le :T1z,u蠲ֹԲC.Rv @Վakȍ`}#D",n^):hb^4Dtb2dWth;_t丝ng#(Z6q*T&#`MF1nk:Xd[d!#H1p*Uf!DCuqULj#g$8Aa3mC$嵔SJXydn jk`PC% Ȝ.j3DQ~3-fd_8DPàɠ<L)E=Ήt̵l-Ã)Q"8L)NioP7e_2c)0ֻj6OsV^lஓ_[1K]3|JNݰ`IY˪zN)Vl?SճoPT=3-.VoѢw@c1!7 ~bHf1$CoG4 7y R ( XAڪXAoQ]|ԵQ4Y[tl+ǀ={g'D4`>ӱ`t԰ibURHod9ڜ9i9ZKnC: T#KAʑP/KW7ndy{ÙSk5Xg}ĩq.j1oQYf:H9s͆ƳD}[͆6]:!ͬ NȀ͂dbeԧ߂F肭:uӴ6\9H `2 @Kn] pж'z#q<9kKP l]: #}O`uh&߳Nzً^m/Dim/L)"l@W;ӫCf!-҇xoYE[7! !AA`"d!ZCEfmZ(]/MPU] eIPb܌ 堊jiהiiyL^?dF 7l! "A*kkʴ,S;[kg`zgv<*jP P q-VP|(6. xLVGc *&V0 mݒfyl>W~~2 I[Cʤ'l,RM)U*O۳i1 VẆ(C(nmNe2iS3#in)=htTVI%G&攴ULKE.5uAʑ4:tm*燡R6]c)ӣs^-ctLDPݿr^"$Leae*F1TAJi~U(O] +UƩ<'P׵he{jVrͶvKZ2L]M>u*dFB9|FbmAߡٹ.}jGx`zaopy㳣x)Li"xkHL&+m}L: M'Q6p{.}{s4g-C6w_?T6l5l6 !&W)Y=RGpqWq&Pu>!Yϐ(7F߾ʿ9y +6__ѯGgG`Kuwc9ק˒W?ky ]gm\aq&*).ל˛>]^|BWuvRPb'gpqyx随 w&۴jF]ODz3#;30QN%p۰Kj[06m|s#)p)73Cp>8G>]ˑv9.Gz9R^HXyyr Hi#rKTZ0ʷbe XUR\3 3DA_y,2(KwVkF[bEZ,3i"qgt>bfYJ4x7+~E蕺a]:ֽzK@4RV:CՃ.`.`|cs|%½-Z `T5}:#\Or0B$7F|ZLFS/WhHu'x0G\i6qsg9G7|eC|/^wz^0uP^r/4kj (CI1SScmLRi1c^Mtȟ9,_݀#h5`}? oO+r>/݂޿гshχA)/P vFzs_*n|Xݽ\!vQx]yIAti_+6NkzߤO}خ7.x],5U(`{RNfϯݦ!n2 wpk#51uFZ(j? Vv^5:lMc%Ju~6\2 r6%$XRe6r5@e)K{u?jex0 utCM31 :Q׾{Fn&^t~Q\LŔd~S1jɒ3stzB ]'>o1b c>15{DNlN@{QXs1A,M5Qۊ?05Ջcٽ*)б3[0Ir9p.hd?~uҍu1}Ŗ^bR^ԿR rޑ)EwiiƨZU/ObB[dj, ۑϮ7KCZYұG-j"fxi~%1vJ@:5_#p 9ʷQ,{?tE/lf\w'yc#ЯFbj[_ 6]L\FH)y 7_ClG6, 3VWN_j LaaR" l n,#9 wyt3FVMLK %~go&.owZP֬X%i{>YR5'ՊJRnɥ4?j(ONMD:h6X {M֚RA|e|ҕi˚e|8ȚTiQcB~)Km[ ԄQٺo&-;fgI+97ͧw$*p"x<[[ CJIh2#R(QQ$-ƄaxfRsHx]) F/K`fG/(/z`҇q]}<Ӿs^ggc͎,|e_ҨИۧ7<7{{'p`}t^ m9R& %T&!=a.(My#*Q W pYn<GdRž+Zbp:NTwVJ-h%TL-ZfbVM&X\;&ƩdXv4B̆@jQ *?}&-nj^0[n+?ck|(vYA?%67 |WDIsx24i8Z|tɋ?3`s؜g_YTZ]I='(4hHAuf@cRE笔i!4qj"̴ά ,l̪RˬIk+ʬXMfR[JYǬfk[Q 56]) jk5жfFi UuH`&HmF֩^?VY{쵂TQwp}O~+fF߁u n{4i[= h{4Z0?QѠԂUj=F)V菨s;0:åF.wCLM meWGlM LGi>2KKy*5 +R<2&0ju Fj%l$6Xvφvf2O7{ ޿S?{IkP,uuȚSWcd67RUwM& p iN2X(6(4PcRZf \sp ֌g&{x>\PoIvy|4ˆy85Ej5SE;]KkNR d,-{&jQC%ˠ5g\Nc&m"-v}c,Lo#1gz Zm(FMՕwiG#F Ϗ ҮffoLc*-ZbjZ/l&en&pi6mNjQqWn`mh454T=`n7 ުcA~iZj+8ݑڇXVBH c(ƬV ±_.yyi__0 O-fNE1! D*BÜ Q>̢bA؟ W{Ί+fSg%QVb&b[4/o$0%剹jjL@Z>,Pw{ya>x1"ϥ?R!/Q>>3+?x-TW,ߙ~+]|+[@ D$ INWhA_8xM-sj5+, f ʊ7ׯbJ7;aOQh4@=3 -F|@ @] bS'H|o^dr\Yx]od7Yx߻l-/&[ w­\ٰG Œ;L_^Mֻ^_WW$Njk$΋g S\C7~+"èXM(^׾G(=[bNz-}@`X(|e,48#?Hm0A"d:UZc*ɔ"x-3JbÄ0|^$?a4JJm#7*~ñc- /|ؗDR% (Wv\硊R~R9$J㏇(Y47R-)E`TRa72|4 Ҽ꼬n?=YC󚊖i"MC0??Pz#/_8gg*w8>qm0dNq_o@'}+JSRN U$gzYr_Sr<$E/Y &c[J= =lIn,vkfEX7E~,÷$㺖|jgt/=Am4{ SρCSp^r qcUnl*-Wx {-`p{JVjn|PK}LR` @BR`ަ-r qcUnd*c L8X:#I`ɣ NƸ7m]@Ȯ "[A|^zU,NRA䰶kP:4WEo*G>VRMybo52GkӝVܡ,% #%h2) +&gŃp̷]ˆ%"]'f;5[HOaIm\ \:H*մi0X0kKqD*jզT94 (ZU^ #Tmqo~ QNz Z&vnGŒ#ݷş{L$Ir܌*:]^]pS\LDp^<Ѻ]7E>v99&vV`~OaS(=ls;K 6+s o xdPiwww/0}";_|ڈXq\D@l`D|jXeT?q!/Z>^ :PCyRIJhr]"rfSݫ{YU3REܭIM,FwxS9GG8*7tU, mȮ8g5pA0"8I <̴Ϥ|j?CL {hj&RYQUρ_0ohNvL9Uώ3CcTs yN|hDM\3{Gy=cǪi~[K΂\-E]v`2.x^I.!3[w 1nǪM7߭ ONs:^se8U1BKewl[1rc$is\[ ~\Ai!fqQOl1Ґ-x>Bzd >qh!r5ٸ`)&)8z۷ Hoz_L[j瑾>?>IYIC% JYADž;8$|(C>Vn:;m#*)unr7耇:l |s͹7`LJPzPJف}yen4UYoV?0wᐶEk|#"Nisn~Cy( 6*c%W<XJONFYc3="\9_:it#}CcAv% : `%.Wd%znCd 4*O/H<)|&Owt!wM{f%]v?ȨTnigWԘR居MgS+L6͗] *P1GA|jrLr\]yL'1[{EgOO#%cr$C3dǢjyÖ`j񮸪X2s/χG;;X(bp%>U nrHHXkˑE Ȇ5a9¡ّ> o{äp#ӕ-A>o&٬"SHNP\AzԱT*||2<3M=OS$Dt!HIve#xtW6|ʵMKK;^VIE :(Jtr!Vz"g9)7Q##H0l`%;oŌG>V ]SgW06Tqwa@Qpq,qp"HFі!xtKPcU.g,wC=2*̔bT F08$sH- ( ,?oČ"7Ǫ/xB5?Y)/KsQ\ Բ{ިZ>_00 fn"8*s1'=2ZgG\Q[_VƨJΓ gƲCim(E>/H5V:Ѣ`9JKl)b(Ͷcͪ)w-7M9t^7^XYRO~wfȠ$T0XMُ+W&k75\p8Z#׵QU;ב5Ѵмa!x0ĽTG璦z=_Q{v& vPĽ=;_1$ hh$ vPUN҄%:zO~xcZliد vA϶@?{ XE˾4 }nOuN116f_}cgD#_Vi@ITTHI.@`њ ?F<B#X/|}ʝNxp5`@>txі!xtPUn<.}g ngW[NI7C :X%} ={*|TT׳vQ銺RR^$c"=2*nϐW :)ג+]N,3^F͆fCUD?lB,R^* x4$m?i""^5"ZJ V{>GeE>VUp>R!pgs]OoVod>!5"\)뙾80P m?>PL |ʐCd7*RbIޭ@rrzdTĄoCd͜ q¤X1ړfPG8JVx}lYrj2W7z kj3̉l KQc-̌e|1\:9c*^4C!)I{"K @ TY[J&veRnE>VsBH^*QtK kP莢0i;2 l+{dl`]Mj^ThKJMG7qXr=`<[C>R &*#my(jכbzdT "Za^TCv0[kEg| p1_dG=}̏Q1xb>r0ڥ#ڥc3aF7)gtGw@}cgt_213􅸴zFwm_uB 3Q3^b-~F1Egt_1N_7{P0!OSngrÞCp?0 qlI!o}TcGjx}>'woy@h 1}ƗL隶໦nߋ5s^'d\iϘ.0{!0÷6qZKı_[Rq(#o$1Ɔq?M&q/K~C;`aV3hFLwϞaq+1Xz У#0⊎?̟env2,\7s(o_w7][uNLVo[eQ]NX5b^/l-X7w5yZ>Ig~o7'ZSſdE`c(_lohcUmJa w_'j+ZMMWU,?-v(.9f|aQ+7:/n֛CYhG03cE PtHDuCi!%IEӧ.r6UvoCC5Uv1r$oz.)Q,aɈmS:D'm5 =2~nb?|ܛ1Bte8VM"A))ijai\Peя}^!f܄wg;/܀V]SuCu_SuL X%[^S e>TScp&!<KT bv"]7!EaFnƉ)kVyxc&>)g"+3MJR҂ tPcS |@=yVRʫFa+amG 6$z-AUV53C~:Ϭ®OQ:<)G¸r]{q :KeC2]WZ-`΋oJTV6EY5Q#e@U|I^;))ӫqe(&o*G>NQ>U,D ն:^qx1&R$߬|s-=̫9'cY17Ao?ÿY HoZD`~.b+qjjkkubNp]UsS+& yBa7% 6-><6;+rv43u1kpkFrjB<6 aG΅v9렯uBȨ>_o/%^娐TCSH[eLrfBH]&V 7&b;}1>|x\M U[u~AcvOhHZ>]]݄s'*:sRa_q;kQ^Q}稜6}yVR!'MF>sZU<>i´Q笜R3Q><0mzdAP:a׉-/W"(/GF^|G*Q_>?(Uχ[_0A"VNN}x S1mYS E>RbFdQoT>EҾBK/WC4jt!uPc:arhUE ~-;mbt(TljpS(bp"þV<& ÷QD(Ǫv>&2 UX#BAҜLA#kKy'…ՀY5#;o6v#o>\[kq=amߥhԗ` U<"ﯸcYڸ  u*^/Ä́IrC4*/ *L5f=_wz;D$ׁ8Մd\)$YJ(=Ӷ{큍=QcUl*~{e*Gy=2jD"{Mq4K$Ƀ5 t"Rq{ 31FU턨'c(Ә&fؑY! +}YoNb~֓&LQ⋂^|7F"ƓRk#`Lin 5OhS<(*Wf6/dߢ4zdԜ&̾zzgN cN?7MPw4z’hlB8Y&F98gEgjXɚ9HAe:]m+ M.!Ů7.^&{F$K+v]QIQ>Ru/O*AtK7e<^|E~v/xO};xŨd~Jh1[K,պ/hCn6mf^{7U3-yp.o!:o %Ɍ$}}˖Pyv1;(sͱu`{ u/" .iڳG$hLt'S2s>#gΨ)y8VJKnBG*`\NhP>|RyqDz8@P E%nW8]iO<*W̼׃At`:NxAHOLγtb5Q]ރUZurΥb1])( t!x>SL[;^DjٽrC E.gndé8..y[1+],1@P.Ǔ;wP[sj.N4}^e2yڻcfn;f>+㪗 pz*fs 3w0p*a/?m#*WGwUԛ^]u+4 `)Kx}cb-;Em gt Z4{s`<<ȔeKm ]~ޅS+<&d8r{PY6cʶpf6?T܁[fI9uiuF<bS߅x,E^Pǩ,Hׄ"9t#}u엨oh{» `bI[Ԩ ӗ|ss?OHq<,%[i&A` Un8^a٩%uEƅ#+T:**Eƅ90 .(Qo6 sZCӻLܔKAcW 9o}} YFoPið~$TqX a ~7)n# ɴ_ _o#"؁g_7B7MpZb) (@R[o/~ aZz=nL$=φbݟ~YhGeUKOUo}=h;_~NR_8s0'I 69a c79+EW[3j𰚖&VL]nԍw,&A]c_wS:oe؟cU= {>f3sCo0|> {}g7dm> Jg:x0hMp<\7eiϓNf9n0ۺXoï~@vE8}~H2q‡ a `<*-ph7?uu={lh+^aem-1ӷ7VKq e`4atyX{U1mc|o ;8πj\jSޅaQ@m޼ n &mE9jIf'g>+ SY2|,F(~k4EoS7'7yH@?noz?J+=ZC?}w7Y]~vah!#eiGqwZ jx]ʍj2Yӫ8M#PM9=7. K뢈۬{A/Q+,W(*lAIpE̐xIѹ.YRG:܆KkL].J$9M 7h!2ԙbB^Fۤ[ i5q6_aR]y١J0m5& ` lӏ6tTPesK\aqќpYQxF ܌yK3iUzˎB\5a1qCQ0 V_lYiY3M(=='Pv32-˵DY(9ҔD>#RFD йv.Omy 1֢M3 Ԣmmx9>-P'ߌkXzM\i&BK1eHy0:OU V(f] Ef#"T?m7,H(7``'LKxIcܛDP`yZ(nnnmFU5>N>r2e錄XZ#I>d_`ʄJA4-ZK ݺH[M䌰<.Y5 a/^VÍf'2^zg%m*j C=!$!C™ 7Fs# q6_1{½.υs A 1;/mK&h\ؽ>h?^u!0X?i5,ž) z  J3M.jG1&Dw2|c0h1OäC!Q{Z~K0>48]-z(rF],- ;/^Q̭Ỵ&.E_gTȨh66YTԺ+-uqdu\2j&v۵'?miK]*浍K箠n.o#;JCBy+ !ƺ+M3p:YYj$,ӥT-6ԋAbEf\ŢF;CuꡕPV[z:z/[tþM[6cYB$1ksxDQFUhTu9nqi5q3cSr"٘yCBxṠ8#a8e0ĝ4R`p+%ϿNC8B[mai5*EQj|) FCK8sXAOc32 : qs h|ћ҂ykSqAg&!|1n)70|Ww1ݻ$YfRy'/Kk(2S.2˄ҖhK '%JRr\ƽ*5)ЋbC1V3&(6͸Ƣ K~QYLv7c"$`*͖{pod o䠎!T_m!O~ҎkS.R:!Ժ=u NP7-OwX95ᾮmLwN<ͤs݅rg!!̳ 8h b[,Qqg*,tt.p]E pI2.4zNckp+J%;MDzqt4t2:E"p8ɭI8, G2Zm\sbޫ .޾ͣؼ^>)`*HDX;QkOd2b sf6.SjZ0Xi56!i1n[9U%C[NoU, >J*̻R0Yf< #CKe^C" ߕW!!HydNP>ikgDߌ<ܘR8/]J(22SМЊ^\F[]S͸E% p_>O5u{qIZAUEq1*J8vu q-Aܽ;MR*N ;Tv.?N?TaP'ߌOUD;XXuU.Tƽݺt4YV/7nZi}ۤ 1)ULNɓúLzNqo{]?z?^g:YX;_A}:ԉQ%w![#I;SV)U"8sIbEr %ay\\jxB͸L5Tt<2쎶|W1ҔL1{5޷@7iIrNh ynU浨w`/ #& y~Qg1,T:ʍnȑB8ZU/?B,'; H(WʓgoEu41_D9/ /=:ڤJXitV, S)8X<.d4 yt}ciE& <]I6vM xW9]mEW0MM|^Š`;Ňk՚klYjSMQH8nΊ:}Bp_w`,g ZvR("ul"ySti(c;w=nH}Y @pd| ~8iny<>?[RD$5;pZdu*XkC:L6;2.[|ҏ/$Qȡ9=}7rOh\tˑ-{=kY. 1jQ.gG]ըnWe=]d{&l(I Mt?fj f`<O*<94w駲5u~6JDIgxPsXؘTR Oξyfv03l`Ћ~uĉΞg+M)nZEY,?X;G*Ilfӣ38!楀0g'ev8>wO uA$2&8%g|zҫ5l QȘt'yCv7~6a2I'EeA:$PƢä+#>?4Sj1.ixuY[eo/]gf;//,w 'mtn?Xj_;ؓc%$+tx(_ˊ`AoJht嗄&{vcn:+-و0G}U;yԕ_}6pú~S?Y]1<B"V*l0TJbJE7or~y\ݸ.~GǙk)fJ-lz}ihPOƟGR7Ƕdb{?hꗹ/fJ_Vcm?דү5aUkDJ,I\44*/kra䈳; k#{jf~nȼo'| k0@]C-q_t1*{?ZK*+Xj?lht泥*Eˑ51Igvr wZL*U}NOUf ><]7MYL`H-rSOz\-?yydck;kG]ݯ2Ĵ<[V}s2ٔ?F0|]SqYgf5l9€U?*;z bC6|60z$~f67?'4IjT󺷜 vkc+&ތ~ ܅|swx %( g٢26>i2hM)༭9羈M;~wo@σ l0ԥ5ҁ?<SZR7J c>G1hv[& UEG!)zFR)aP_abJ] q  ްuV, =,)G")oMy:C]=< Hv?(Jh8$.2<c`6@"4Tj'20󃒮RR07PaXuP= p*yE 89x;+U&ݪ2 hfkW\"#j9 khEĈ!@> @>)7YH +Y.,X0j?c[Ud;RWJdR<Gvus]`!s ќ,RXfƸ ^].ZID~-2!n,f1ux5ݠ q³('/댥kIϼN'e9 Ҵu˨!&;!D,u\#B:&x_ pƉAQDQW@\'.4q.P)b |\B[7'uuFB'.eazͶJ?PaӯlsKp3$7LُͯCn8I"?) jO_n>_}0YA)jI~IBf;i]kNv 6]nbpSbpwhbpb%%vO#b>1Ny<*?N/ŝU`]CP`%%n+L5[mtSF7#^͆651łѭ?9urot "/x+ &@eg Sc (#6Z:H H)7$)v^  (I N @Grn"zT͔J">9Djdg @$0_YM}m30@xB=x& hPoev[ <[;T4Q1La6P(}m;zg%ߦmg k(GV|8xk a4;g!4%PwIPw u8T?[Z'Ο#Ǧ.esf}JvXZgcEl90 J S2,r. oMNA]Dpa:b \y~maG0Igxan5@y qF/Eqi)7os|-NoL[K;]*L$̮Q!ajuGMHE n_9&iK xA N)ZdR#E=uv{}8{?GóҳMeh&n)ݣ+s_cQn,{cp*xN;LaIIiVm10ƗƂ*mX"wB\8M+T 8ȸHq] qlVxK xK)ǯ a"m(Fh րhbp))5* '6 '89fci4PdK84 S9(@A?gdC&!4B7'_>ܠn$"/+<q|S >7f6*p~$ջx[8>r_B| 0JA6]( (<,UH'CIOWb b)5, tSهi-#-VClI7Tpf&A8XCk1[;.d_ZS/u_Bl 56zn<31L Հdi,U,d\䉴+v[}k>_Z[fM]=D"_vFrDB?0*R^ayp ڛi:W? 33x B\)oD#T` VR*€y"8j~n$Ѻ-){'C,Rvc3-2J(?:}1$/IO0'5q1eB~\od~Vc$!UT{~B>%$ci^x2aNN0: ("X:(=TJc$[>T sa;%)eR6l9/$l`P#I=3Od4B#@1aA3xBX޾Uu0.""xMfau6'WyÞW(3'5[ aqL9]YWK zm8c+JT.j$vʨw{!iǁ)-D|sǁQIRݘ.wS-R荥0;{lReO:~I1a0)N:m[t>B r54EIK+.C{_9 A ­sT:5Tb`&(#3u E\hr..jo %8<@ P*cm.Cnl.mשL ̑|ewOFNKE6 B:xW@#,ߡvR8|a0f evƱ C]6ʔG =cAޑ*ƭ/,Zr6iy{īa].2EEy4{`rxXK.3|pTIR.Dʧ[ȨZưN:·01iJƔtOu1yG,Y9/*gN|0+70 Q}Vxlmj?zVFWDyM.|#[WS>.Zm/h>/ ͽN3@!WTl]"4@i)0aqayBlRŶx ⻪h%4,!vt[Ɖ*(Ͼc uv |11Cj9 tL|c[@*Z:H,3$_F(n6tx}u0שHBJk>DBBnKFVM,uC*X+Aiu lt_^*6KƆ,^`MpalX) UeWCS%҅QD 8cQas'6H븾hkj}\Y8, ILł瑱SBS]hR1q`T}QIf{ \DX0z Yb _zd8d_܂<я/`( \`A@}p,3`OW9hӧLJꮅ[Km/P{bGi:v ;Eai%E? }Gؖñnhtxxެ0'B aOE61g!.pBM| -lK<-c9R IXBbЖ;f5hK_BK%"=Qqrf[%A  ׂbp|wS=#M9A+E8 n?hh"_43JcU>)Q9F HIר>mv@*A*-4-TEBVSAʉH.H-/ZR0(1hDrmqĄc\)SVvnDЍ1=>gpm8?ݢK-%gE /j$ "ijش>k.Ѧ\')E#;*w4n4ԍSՉ{gJ~SD<{V9N8zL$2C#ΗMTR7On|c42QEF}ˑ"3#R[z QmG4IHkQ~%i:Q2QcYT I=}zj@';dLE%8{, XąV`%Px~jSP>EC(;xTkȜRrE Hg .+B:V T :ŷ13?(ѲU |)4VKAJBָAHyQz+hd'Ac^ΝxQiR=9!Zz2t<٭1bt$P<浪bOyQ3cPc?<'P+DE:IMo]ϧ-RpySҠJ C3M{Y [é|1x#E;[E7emϬm}:qJp[SQ)1iŇg[]^he۬qӺ a!ZK_ 4w`U5xOKK:yඍۮS 1׌mQFh8o@u$2C - n~]y ܶ$FwDH/j3&a &" VtHC2aX~_Y_ Lv9>9)",F^pDнvC;+F*U־heƫ$),z6`}|]$?H7Oi D&6͐C]7 Y]GQILwj2!at@!] __(Ө>T&p]:}2XOM.o[aaѹݥBKGcIV N EbOz|n5vWN&j3k5.dz> iD|ɊC(KB3?Ii He '9ڱqhe2KZE!36e iqKԉ+hd>G#*uCֺ3Eu31DדiXr=?n{HG؍/p$lh۴F3vf0$.=5:Gc *IUh8~ʐ~~ 1nVՄk~xضUϩ0Iu9C?|fH v*;o8^[ώ@h=;;P2clLᘃV $)/>w!6%AC"S".Qwe NܺRmswpki| ]Vr\x b$% -\N-deh<]VI_h*iaJRU*oغsWIȣ=dj 4OON++hdo9,/S{V N{p5.hQ-m!t3h JwW <3, q;kaB \ _~( i{A#|߄X N"!Jh3!duVVuߏ!x(|kԖ ;8Ě,!%y! b7#-[~?tf(FН6KKμ(ӯ8JڂGNF&E/u>Q aU`lr^YçF 'qyfh7Utb{F{[X/z{PZ%nO)P?T.'ӛ&/^(pϥ)(=3HF;VzF I} Inm o;ZJ c m?Mf?lӾ|Vj8v/4}y7v է߀SsWXwrb/guraT%ma}'%2ZJː f%-޺Ri7aoR/ ǿݼmsWIOFj:Gޟ {j0x<'gn|'nܞLtr>3DڏFs9s2pf罳|OOv5{7 \9{y>5eFA[KaQ՚ hnJH{ml͎;^g%śe.|]Pxd^+ 6iqlROdy{W_5_3젿:}v6>ѹ$/xo&gh1{n?(E)zm0fF@|uzL{iA A0^mm'nIJ]Y\!K)_Yw8TtJT A}~AYrC})L?g5ڂ% !=k%+q]Yn~sh{p$:$I\Vf5T+潢2( q[y-`U?:wK2[C U aJzS*gDq54c<:W{7ʖҺ/F4#r/tD2oGGAÃfϟ! ~p,=V)EXYAs cJqLL DG DSqe0MwDŽbg/AVRl?\PzFǕYv{N"*/B:F0yA2Yk93h16 ʫJFU 68q(5Uʔ?ޫ֛#D;y,ُ}(خv[7G-`vX x=X1 G$ZR .f3ر4_7`- (Ѱ cFVs2v;GoSpB0iJbA{a.D %U#R,8ìF 5_ۿlS \ .~·[__?:_G7H%n&7-1hHʀjf(1J Die>\L?UE$h^S Jz xF5%2X!qT7o" ‰t"0h\ "0s䃃>R$hjaܝ?S"cS)@A- Ā[ QJK ƚ> E$VR ƣzyqA(fFv:nֻnr7/FYnZ./@ (r3Zv}O _ߟo^Ù;szewx6};_—>fz^Q}}'=z׍3w6} QB=f{p׍5—Gvr>n~MN&:h0}eǣOm45qmdNH @MҜïun^@|o&ë\x髟]ϐeZ_ ݍ04 0TF~؅c|]jeݡw-RnSn1 oh.d|{aVSpXz1/ظ?`7+q]W˵-(QVp;9[Vq=v}̇坻}Yȍpy_i0&èĿW<8>/"@PxmϜL2l]rԼw.\o<{\΍{-Su/H7_a#N=!8=7pz/{đʔ6 m*[,8*ĤVk8ɲ,k&RBHٖ:_yvū8><è0?81X,% i=IJ:0٬.&=WL^G)]j3W譮|?=8Ë+ٻ^CD.̳Ñ!J}#~d߀矫x`V.h,i^-g4&E_SZDPQbpQJ4D\Dtux^%,Fg'bw⬣'Nziq,8ܙ.kK!R;FbVʂ0Q\|A$gBS>u,yHtykGt6)8 z&ig:J[괒~ع4K8|zμAA긮79-G[΢Rz1D???Ng'GzY +)1|co27F!MQ}d)~HM{~Hq;ǵ;XtCpMn鯢NNoS i "+K>ؑuM7`_pM5wOgT i6?O\]:<N 7Zzۢ}< }5իQ,Y\Rvؚht bއ?Y֦(kkkkke_DY1ByB-.l}|Yʚ{EN4GNDLJQB8Y2FOؕ$.(TZ.sʽ];)صTI\ m*|jFhu].go>k7L?j|`O? BA1_$ P{QuwJL*VquRYx4Vf`hA+gi?%_Un[/51m)%4SDH/k!+blqtxR=!T9A͜l=G_} )طww~ͱ Z|=h޻^2?Dz眚t:؃@$xBH,J /{Q۽ݩPSuKN/C8}~8Š10yskb?AquM{My5?W`?E,>O l0s†~%Ukڷnkk{2<-jaW?=л6·q֛?=;zzr-nNF%rE}easUnFsV.gnܠ>% } .__BR/?~ XhشCM?UuwNm^75}ٮz?faIEX(mbi M,' aK[{e;ˎƜA{~N6d\y-=q`XyXן0'_/xK<%DF8/.}4Gq02%ŕ. ~Ɍb- 0,P GΦܚlk _+##)6<8$YIwV#c)GY5vʶVI~eSxFη6wvR3/z]eVC{sb a`aJT" j{y_x4i>G1?kDv8Nc0(Yib]1Ya.gW°P6Za"4aVO"8;+­{Ϣ?=k%̴+<{,pl'kwNj?]Ҹ@1>x|8. tYO>RbXuJ8>dkݦ8&u%^~ll{QQFHx9 *HjSQes=&h0H y0Ux-_-i.U&ct߽.kE[3nV`0 9ʔ1\*"'A@Cv-+tC++i $  }D'7*&εJ Jw3Q#S-h(+4}_z{ !LْSѠ\Z.}n`7t1anKި"~Cn걥Xe][Pn.Cz+]*>!]uNfiVau٣qozٿFw1?B+Di)ي,!And|i>sxCooOܥ'8,U&\WdϯQm#;`|q!P&LX.^l hO^\pN+U$ˎa崎)vo6Kī ,B Mb=pb晶88uyӹ୍R _rqǽry <5ET&Ӱf+cBW ƺ-Ie >W9 R{Crڛjk+{sg,dBܦ~razdzR}nH ,oTxڞgti7vnG3C2Êqʮ\>I=D21@1GiE/=`b A Ms01 Ro 1%:S.{;ߡv rK/C9X7! 㴲nN!xYvPn| 9R4X#-E |; 5+ZvOB1iHI:+ 2K| 9Sw)0!@t {M Q(I~ @!xLnGÁ2f#+@yWNK| 9KCMnMX] (M ̴U| 9<DRk%4b9>,-ھ 6eK(otf kqȔLTm3.7:\J&A`\ZT.iVҠZd $>ҥ'XPȀ}F`*k1u](5W{I Of1(l̅/ |;] |,A,"t:W"' ,zQS^SA, tb2UE , m2߁B>3xA/RL?Ӓ sڊ-aѸŷB!xsdA^qځ b`,Aҷ.r$*9%4jrB i߁BJeg{u F6"4Cn;Z^ 9=etr> dA0<ŷBawcr˼Lj*XB J`A]i8XA[V 4}{ y4Z) :JnSmA¥ | 9Se̔ࢸ)AcxbB!xp69\QY¿4 Jmzcji?db-ЁB3 $%LAu~[4[Lv ?,* `20yڸi ASe %,;x1qqK$2 f]i m)d/pv7>FQ1+DZnP^ӕ ĉUN*)𰂙fBc}@ Yh5uIm1#[B Oȩ | 9%s$WiH uN}| 6EW" =*#@raCw<23 AB>@iHJcpL)ύ-;PKX7WR\ڒ(OZؾ =<ˈhiTI[;gJow"N烥UK؅B\ܻN2B #Kq4 J[lv<8x$;x@5EDctZ)ՐT%CUu{Q1syKBs^+XN9P^r]e 4]}2%a؃rY $~\I`BbHWT x%B!x g1QJ \hJρi=CBLB!xC_ [ % 9rABX"de | Y+.O VAl[d )V2̍F$C/13\{J .-;PJ7#e*%*`D/"!%B! xs|4XbTZeF"G$lۅ)l^zdRMO4"X-̙r<֕[Ktj~.äDa**p\U k!IzOPcXaNʲSD; ~'Slwۛߟ=iV`OlTmUh~8 HW%MBNodpHeԴJ9Ϊ'Y0rN UJdRGfakOUfak6x8Mvd[NxM|\])wqDtny8z8{ӋQ10Rxc`'O]gAgQo؞O*ڛ R$/^IWSkQx8ߪ}XՍ(j'?U9jiQJIfT~|X<8~dޘ UʥR2JƋV`hOC?ZYzf֝i0bQ a6 !ܫ 4 M]LӦaŋ S_ ?OmR藪UeQ뽹BI|wղҹ Hx[j_KRYaoGqt v&ePIf-ZM8ЯWeG7N;Q5'W:^0DzAb..ݾܢqõ)HdMa+W~-͋7 cU]|c_ӏJ->Ɩ* #[r\[L ri2(ܬ?|e1bW-D?Z֑C:ܮ3RK-cT2YHZ#.v8&ӭRjK'sU_*;i3$hoeu4؃XԦ_/TyQmu):f%UzCٰkOk ZX Ѥ*CdCEO/D)B E*S=^bxsv u >:D! ur6hnX◶::Hx"Tji/NzU2d]o M9߾nn4ARVE$oֽuN4dhրs(r=ʜ%x&낦YKû)D$6)hK 9ƯIݯ2_' H/"m6:%ݦ傪a$ U|i!`ֶڔnWPf'j_,>YCM>Ty'{X@榤Q1!Kuتg](tGhOMw_KͲoGTfɣ_f D Uԛ ָhP)iE +1ؽ*؀d@= `-iЮcSsmF*[ $jɭZ<ĪuUr%,4RN:{,8քFnu+y(!Ⱥj@ 5Ȇf;uF K05eBE@Bo4VXݘu膵vX*P*JYy5 eogU T)qiilP`k۪nWH+5+0uDq4Ս@i,Q ꛐIv4^ DeΦFcAo+`WV"w\ՠ7wVrEŸ֍@CMc5BɂA mG!HMwآDYUoYy<@Ѝ"gQGV nB?% 17SL3H L7њ` ,Ql6Pr3إ~.J-+T3[!(P.hqA(xUQ# 0}IzUͶC) ^i;+ъacZ#q[E]%E/[.R˴KȼBիf!=-fkҠ6BDj&tY\D "omXE񮀵Ck􄮌I xfp7AY3k/HW̺hN$!ƘP͋BЎ&I!Ԅ6^gUv]8!\,MPs.x v ^ޅJ}6=DT-do `9:pԥu@xc#PvYd :JrQJ%V 0 hd7k=ȗ٪q=h5'@i4\@23yUA>xmB&-J7kE?/AXEYь6YK1e@TDfcU0~~s]Yz 蓥d}E %XVGXk-]EPhc)͚.O9 * 1hxD.c`jYwk)d*H5"%0ZuI(wa>L% _3&c V1Mh#{Nw=ZupP*"tQ8ft]Ƣ(VLMF"Biv%&@?xP*ZE!7)p*y`*leI>fl43;Z@'J sAn+ V= w64IxFDi3+jt뭊&iYFc-A*mU(am W|ZR0 1dkZzv<_m0, Uxmt>k53I^U5*pqt*D5IҢGJ$vw4Nd6QY^k Q^yhY=y`4x&p31i` ͣa#ʷ*̈=Xek8nHJxKt]ж+9P~rڪ.-H[MJtENԃ ( hT]ڢ)RC\cn#2lbQ*'~+ MISL1. xrSpm$r,ZV/_*IP(EcDMnƂGtvsPuƢZ'`ҕȪZ:hUD̩m X{uAfa6#5i b $_T4/Q40-lF]΁?њo9yٳNK^ vK&f(֫YëbmQ Obj* (FWP ̀~jeRt_H O30% 7aFE9>X2'ʭrO j-l`MCɵ#iʭ_;^#p(`"n6j̦rѨUrid" J.YxY 4IO . $费tbm78lIҨ5F,?!t+zgjp!GTQlM~0˹y>_)p-X:Ǧ=N8n;/ }ݕ&t-ލS21t{oǝ GC֐b=n4/x 29\I4ז_FS{4c hG `wѝM|5G;ƣVZl9ڹgEHVI 4`Oe`"8YMB!aP-'e*7pVF{Pp- ׎_rŽHŖt/Kaġˏ~IA91)Ůnyc߻`<*_sP\9)]87?X.a%{JV;Z39Kܚ\vlr]νO7?~x!翘.& h4 'gO'gvnM_lz;|b/o????gy<@LA_" Ͼy=X5;y*wI򨕻Z{aŷmx.F}> W|V@1l8~?xmX<NW=.i Kdtr"hOG'—HLk>gQ! V|5y:KizHvMÆza2vWPw=;x>Wj>C+a~ۃϣ$&4X :ŢW<> ]n 2o~0hc/`搢å5![)?dh5z&c"!DK? w&E&!买fbQ5 77i?ߞV·X?1hE^\U5?E߯`FzK+q]:eM F0~|O?deX-j؏qr`r>caGh^_µhr*^7IjXt-7ipM4+ul();dBxW8o}&~)_)tipwG1{b텏 |~G%x 7A9v&$Q%|o7r-_ ݍCqw2υZF ]P·k~ 'y07!3E_ :i6|4fK zvvW{Vg5n7gݦ~ o;-Rx |VǴMt!z Jwv& ʳחw8poZyO3~~}p{uV{#|%o/Ѳpql4یxV l8AW̴UѦ8$#;xBWz%~kg?7B nnˋZA;B҂r/PИ@adӾa_4[hL3kn\]Bh^GdyC-!3yrєć!.vͯE [ ^P? yqV`dV^MQ\]^7j_Ϯףoa] I >^68-G*bIrk굻pn Ryirm|fm I@4;miF5EEވ'+!T? ^V"Zm  j@P@{yvU8 $WhP5e #hihLIVa47 gV`i~{sŝn'}S~^T/rume oGqob{nzo] ۟?t⽹Qn@-=|g4qƉ&fdzʝ_vf/gjGKۚ䏆8nm|/<:Wa{B=NiD;NcXORݑO*׳/NF 4>׸PnL~ a;8 +{gW:n /)HIMV3Hg'5GnU?>ac?&y:z?7OESc? 6ћe^,;vߴ]Rhţ#$Qe/gSjDxә>xS;*wx+|eԅEtL0JiLx$v5EOPn*mO*_cRd:-XE}Xv}7~2f_ǡ&.xY4&6R(`o{x4NdRe ɓğ^0B_ZxVhi,.#2zhτt휧 ."mŃBI1P8XC]Bo[^P# 0N%tTmZbb_<`%J.1E Ǹ4~UUGUZ3A fv0Y4~qWPtYYb.ˡƙ岬x2bϽ„K8~5Jt}XFCj)ug⡔p `ͬINpgKr@K*{rDIKNp 8 l49 Zfie2ܡ5?d쵾3zd|#Ӽnsodͮ+K;'/(/Fsg<^fޓ3~pVb1F=\ꗶa~2R=3g,I|dtT3􌢟h}g-6lýA/avJU5:QբyT߶˷dpLQ (TOL%ZaO ?|`Y.J6XiyC;)WW;kosug% *%0ՙ>iJ;%XI4픘SbN9%攘SbN9%>\'\?t\?f0w!Iw!]Hvz]Hv݅dw!e6B|B촢.$ G(Q0)r8\r KrYp\tJaQLxlgMLx+-1e /('G^PMxrƨ {^8.h8d?y@^TJXk]c eww!J]4`1ÇAAk_`'p5S0p'*"[UuhjWA6IIK55,L\ ;DK鏺a|q}8GX|a[U0/vꎔ z]^:0&¿PaQMQH!`b$2'\!Z..t٨Q)A 㫥z{N-%FuAvq>MܚwLVJ+Vq~%fK}&,ךNQߧ Y]L*TA(W3th2GhlRԇcK|\}eȎ3y8\.yb✽FX-эM[%mDahs84.XY4l`0ȠVV/b;&H€LH8Џ*L$U<]]F}d$s?/+؊?I.7^^d%gŻ**MYEI(s*}B=CXN"Ժټ%U\NQYORizm-e7w{ [8 U)jǕ\_ݴJ?v ]\*5;*SD3Cgi [hE6N+5:^o a8kC" Z4#\-?x z(EV&+D+ġN V(t&oDGKL&3m3S"b&%=20$La}2-/TI͹5 BfP{Nh"\Al9h!PR"ܱ#<`ʦSL8ytpg ]L NWRCW"e׋@>_]e}{K&|GDU6^] t%]mҰ8B 2jۭ7x8<~ qTsex(*[*H n2_G(T0r3}xee&quc;JE"//~H\/ N;90Eĉ}QԪ㍙Fbʍ=0M0Yd`%5 ݩ+Eh,E(5΢9By\Ic]`++-th(;]JC#+Ss kb+[ SvB9:Bxۖڞ-O1SE+D)1ҕYt9QeB+DͥT T πǭ :Oj<SUpi~_5yi޹aۚq׵vϦdl1"w4¤ͳdDyцkU-2 eKJ^x7&I5f3U;j[lMlڞ8[ōqyɕy]an(KSVp(JD笣ZGYsgK1*I9*G:ʵō^K.tRsC)&55+hR_Db796#ektN+ d*PqbXӅLJLǕ\iە~.tKf;&cHrc(>nd -*!M8f4EĞRJ&h[NaujtvZ蕆}:ij7d)*vIjt=]Ξvr*\d-d@zCWWo ާ4] ]Y)6&h}0(  EoGD`\DۡU{ZǶ]فv|\߯siױ[\ZK"e/ (;c>!=Lٛ;{8TrуoCMDp%AZ<e ͳD4L**Gt:˼I ZaNWot,t-Y$E+1xכ_L4V_ћ.Ɨ#[΃O&gm&mv=ءVUsW%Uߩ5˳bZJN_J]J|6.Uӥnny{<}t(6O'Һ=}EV6uy&,bDB4>G~T')P%hƣB|-M1 ObgF%VFT')s`޿T|zm0`Yy *(wt|,O/&,Èg?4beY: Pw]'8'(n~9)LOմ.~ j*njZ{KmpRLxgyi0UzC 7ʁŔ})י7 lrq~gU%Ǘnq q0U3O8uqon>H9Yr51*O',q/21Eۮl 47xuzTmS<}x^5J&b@H`5n2A*fOAm< _F,[zZU?Pl'ǔ#zbP5X$NJw`σEN4^VǴH34<FD];0{ tC+t^;+I,;h[RCՓhz:[~JXg0"2/!:cB! Ը0RBBR)"]"3g nW=¯!z4h/Cu{WQib 2@}ZiS qP %D_a]zjqVws(wqGC,MiY&3. 4af\@.'ywiyYLhZv=R׭ԍn%Qzc5vw1|/#QovׯW|ƾf2dY>=UjOϗy&çPL(҉f Da" ~F*&M붞t.%eKs҄dRFLPL !cFbQk3Ek Vs]Nz;nn xp sedrgNuȢ=6O٧Zd>ݍ1;4\{ԁwfRpCC ExqBE [je2xwxwk#].}PIU$BIhea6њ3LI\F"#*EĜeNPݮZyup_eN/73AI;pi\ϟC٧?dWګC^ &!0lb5dŊW/lQ0+ ?nP>h@ɽCC JU,Ht%N*5ie82D%<Y&"I Ō'FXBͳdYqW^ʈ?'ӤB[˟#7>ˊհ]3mČg͔QfJ˨/)-}L(Ű7S2Wt xCWVg7t*s_,g` u-_8TR+(T< c"LؑԐH;sz|tEOP\wb0\*vnyctzdJȥ,ZVXZjSu hߞ4.uT/iC:OH | _?*Fr܋!|+h2K:Z>u,iq r\[zTL1*'7SBFi<*Yշ) F#%0ף]i7u0tJ>ήT YᯃjE5R,k~K@g#Od1F%c.6>\4 7-]@XyYi5B)F|1+/{DD9$;~,{clž{-{*F͏,v6 *z:O$gf="6Kca=mܒ*5:5!5˟QCYFxjvjz1ʺOW[}5tw6܇i3*M[q(ʴ^>Lvzo-' Z -84}U %04,:H-c:K~6V6J9Ы2sH!PU<յ󺶿PN&^7O좱?ZA{9Ky^"e# 娛d%'&7X* ߳<ۣGn>.ug )Ÿ0^g!w­LG9} gYuSnӻ">*ӆ L*xB`[e_15Ǫ)1^`Qپg/CR0 ]!\ ]!ڧ;]!J:@TǻB vDWV7ѕRZ*=%^T^BPͅO3'ƛ }9mRa<f0<+Yp "zNW؝C+ꩍGtchWtp7y Dkt Pxi>@zCWןԈVM]]ezCWص#BO NW !ҕe ZN•B%%CW=x3=ڠ&ӊv(yϮd- t)J3$ ]!\A}+D+{OWoC t,tŸ'bܐlWZ_ jwBvC+.R>LyCWWyCWPwB tut%BV•BC+)#Bʟ`jo Zw C+%yTtzCW{N+BоbR `y3p7th;]!J):@2RSkrWoADU Qtuteԧ`JC.eL>\o_]#%_OlWU;jODۡ=ۄ[j{у5^߶ YtFi=dM6[-TVS:GWlP_J{iנ\o\y)"֛}jVY9D40ͳnڕ%B>pZNWR&\ Zü+)L޻]] -6Ÿ`jo+@t(9Jj݇366>DW֐dd]"])C)'wp%&ՎhUA@ "Ab}J`A7tp=+Dk{ Je@WHWh.Gt?BK`] ]YK0 J ]!ZMNW ^]#brt ާ {NBɞrWP-d-J t0ZI%Sf+f#VRo j ]1k}+DJ)TB v+Ug`|BR]"]YbO+y]!ZzOWv6@W/">~4:,bvR,e^W{ LT9Zq6-UӕRo'{7IA_@׎+* RB*% n:W,"?d&u'EJ ![Q>An*J pEUZU3A &5.m/l8SVa$&1骊MZ5w׼}oepUg`)yEAٛ]*%U|+0Vߨ+(Zh k{a8@i |q?{9S>"5-x{CG9i7}gׇ-0I/g<#l|1N":&Lfb9&K2؟W:аFPuMҩb .'bhMC n./o ji j+0ҰLpJ@uM+5RšDiY{njcwX̘aרfhO[3RzL-x# |5zyo[@ATR4pFߝ,j2 աL6(O@>R;&ƈa4BqU7*)$*G@ȨLҾW}}kU 1WmQ TP,Jv>Bn $J!wӧ$ͫzM5j Jj MQ Ar#钍5u}gtsaM nN'Q;Yn)Ǟ\1c46+fLȨ*OM΄|{  AF5KP!RXuݷn\ D@K$X'/V[ZQW!5*dlzaCKU+ܜ-QAQi+JO@ezm|#c9MF3:yA Av%4/z3|W4G el8duck!$ ( **ڕ"M4dD5 <l34~JoޛZ│D]`)MA)udIwB=A/:j:Cuw Zq*NR#WfVˬ J*H{m>YcIk0 !1T(`5Z>1X3' Y.klHľ=(aL! ∽}Ai!}p#ePy@(ԥj*٠d1fPTԃ,'$2C8`L|c]/nv]:_5|TUz&L#&V&3x:p4oIavl+MTm56b U&5:DϯCO`]0ui {=dHiD0iyM*}b1; ;]Ks4/~OyLPBb2$kuk+<o 1t|C`¬JrCk4=Vdxe&[_+Hjҧa1\\|_u'kWمX#W^`>b=ѠeTA!䦯Ay1ρ 7DR&/ ā 0| $$_L)X1z. mM!g8Z.fX;6@!zjA jhjw ,hͪ]Fp2zvhu< 30 td+%x݆i63+I@̴LZUה!T jPk {;z3`uVH 3v55KA2ZuMռC:b8;k\#YA pf5ڀJ5񭷦:x$`$XHDIؔM@[/l5S|cȵ6vٗi`;|uwpvwwٙd'6 PwC< pff=+[{ (XQm> >;Z,fW:5WZsLڌaγF9@a[7n(\d~z6gidE yֵؓJC7v#a?T&:Y%\aۊNv0 )aTХHπOt4<<X7oc3옍}u`V&5sJmhOI#GXѡϺCE7L ߡ0mQHitrQlFR<z U *cKULmyLG[19mp@୭XEѴR;k֮UVm(|Ϥ@L@rZx%==jyhdx57לAPK}4 bTs(b*f]`)o"0hlp/FWfiq0p(gf(8qK~ !nI2[1; i4 6fs˥ZpUt.!bc-ݱP| ~gn%vM7[hoVoҮ₟ͫ%={{)>oj`B=3Yf'w<Mx&A4?W~psVR;}^fzqCOqAAi91hR:@K>{P() `{L*wb;^g?n.W/z+fbk{*z>2D 7IIHE(D$$Q$"IIHE(D$$Q$"IIHE(D$$Q$"IIHE(D$$Q$"IIHE(D$$Q$"IIHE(D$,Qt,N9p>YuDВ?DP:9$! .A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu`ܒYRPr:K~1A:@uN1OI:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A x'guڸкxA[C9ŠNPpXԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A HPG:ԑu$#A MPOC|`~YtO5~}bc~sDF#%劂I.'WV/&W>WI:\% `.=WvtŷyݒMX ]1Z㎝$K+CW 7. hRNWR ]u}Իgʣ[N7 wӾV6jzۑW(|ys՞anw8V]mPoݭlf?oW_jվv$?AKb-/~[K>bO;fj+@v_U̿vί՛}зs_%1_ 'ߞfMVu 5F]Oy|${B&L{ nfYZ[k˚rXsN;V QXawQouthTri/UV /R1owF'xҺS|L =>|I?y`z%ahG0qp4NNfEW Rzwt(:A2I`DWtZ ]1\BW"mӡ+HiZ]1bq1tZc+Fi !DW86z1tpYLJ Gw(S:AJDWLn9tp_ ]t(:A ƴ$u_Y~սGM*60N{`sgb (uutvjI20జnV;wס+S)mI'ap֩'jc3t兮>w Bg㠏n K+;]1JㅮNNؓ^ ]1\BW6c+FE]"]YcNn1tF-F]1ZsfQZQWHWdUaAtb>uah}qW/_ΝACW6c+o:箄 ]K, q1tpb ZENWh%AO~dQ)U).7z9w8Zk~Q$tutM6zC_oW?7տ}ߞw޻p|soח4~y<{WO~jw\>.zo>.ח}73x~ejSj>гgCmox~i9;ff_nS_껯w}CN{q@18ܝhP7)9O}8 y__u?Β 7hO9.p4m#E0!)\hA^^^ZgڵdZۛE7%!7CryQy=~bv0W!eK r}7:3l-/VMwmE ^.Zuvi;Un]hj5BX^/~t\jEyr09IRN0w%K>wgA>)ϛcfڱU~ޗ-ёI~Q9*+u{t[Č )Ɓ/@kJ IGe6)nph#ͼ?%1|q"~14.AWT |Py1)ah;4vt>t%N?raŞ:~)|)7o|TϽ9Vf5 so-U 'or\+,P>0*Xh]zs2}LJg՗=X$9YXBGZfU&oPd29+0ÇJ0z_4pF"0سdh&4%Z&YEL:Ta,JK2xjԩyA1&PhfUdvtѓxV<*hw4|@)(}Q#|dK]'II2ʭ]BL5N_)Zq_aPxrZ5cP)Lcn0ѳu\͘͏[h1t2,V q.OCg V7:k=k}ѵrsȷF}ޭFAkawbƊb׻VP4wVW=jBZg(E8eÜtY=DJ hCRdz3NU9]XEOvP1kKM4@U`8 ,dqK+)   ̪=tmW+nnZ1<}e7x vu_y$?W3|1] S2ccSmO`ޥNmӳU cV7w[Z,6Tc(x]+y;?Ws>4d?bzLjW{PԄvU~2[Q] lMG!beF j0a^83y#4)y܆В0b2FeD0K E&*#%X:F: -@W]YSp]M:t ˮVQMV[7ÛL4WR( ͝OMO+3I ;*d2i`sL^sf) ',$\yj3a)^aP p)ON&`'gθVCz}X2Q<Yi7w'ى.jb _oXj^k)VSqQmf ^F|p"w# ܉ ǹz7* yZ^.뽇喖Ƨ'ӁTLkJt `+ ٠a^]ޠ'A 41rÁL%`? s来$h76ݠ 6o4z~/1(5y.RJ7Co&۳~ Z֐w4OYV^9h}"7~Mz%3dTgһnjaΙdD)ӝ[BC"RӀU"{,zD>I2,:V0uQm?Bq2gɟuM-cJ佹F?G{ ]DG`G#L?CDF)v_?5ëbtnHV`mF{2ÅEBFr‚uW=:R289w~i4v诒3].$S.g:<<-J`qac+Q տҵ(`Ƶ]O+N@JUdOIU*(`b|x5cthsp:9[[%Vglk ZD͑0hlc_!Q壯cJ dⷪA^Nw?~ox1&^wq[_bgO~>"w0Wɉ._,g` 6(w}z?} NheL1c~]- m@;4-jڛ5M%[iKrK_m7"yԕK q? o{jqXaW1[S(p18o@ vmcfgQE;$%V(C6.Nb"忄mu$` ,\ƨxG"spl#8K퍺V´_c\+)t1iIF,8Мi; SA+ZoBu u&[Q;p;?|JZD-$:\NOGyl'*U ,fjFy1ļ]_.CVIrr1o,N-杍G)eC̦Lsf0M6~ NC7bo; F&j"n2T,@GErylQTٵҍZ=J+z1@\:Z@U;6(1E1ydi 'qO/.PYR|_+~~!+Ū@؟a4N=f!e` ‰F w,zDd׍?uo$BGӶ\He %v+ۚ"u&d:@xؒjk7\6jVꯦU3 'Kp]{P\Io)NsEQ@UcOry!`u)3NA? \ %+Hў[]'-}g;q#ȡ~D#-dcʶe54|lFb`}l6R):s0k wVA^m!(V)3&Y.͢[,24 k̾CO ESϘnZa!mz׀0ː;ޙ\Aw%|mV=@%A_ !m:l!H K*vFrH Ԫ{4RI~XKW/PE=gmxu$,gw ()uWxtqQXm}oʹq>QGs\> |DDƲYof9DzbI#N(ƎaBu4=ب#vG!6OG ҴŔ Cox%08;/Q74Iw78U:EF!O/t0O$?[P% &1F2)Kt ۥ:+vtv-uZq5YOM:lgZԥsU~N Qؖ~8Nťf+(R2O!5AYI>qmD1bm>}\勫NGA ZrAƶF E)ˤMr:$0 ^>kώdh[yaC7((3逝TSe.QĘQH'ți0ˆ.,#|XuN(H0Qьx-ҡ"u@@af5K Вpb k*(BVAIo/SLOc!S Kit:R i Ljqd>H}GD|-{uWm"&Z{X8Z'2K3AZΧ]>%NTJ-hi'Ra5^0Y@#hLKd7X8O)aLGKH"]g_w'ѣFRߨJ{ӈxɭ!wAʥ)( Caԡ c!UZʔ1SNM8 Iu6:쏇6QOzM$4IFaX&`a'/'l*]R[8n-^[Erԁf6::]vVߺV)kopV)R= s~E\75*K4ehCRdz3XoUNvuѓ>Ynzx.w FóU7$䁑c~ ]3|1]3Dݺ~?w[ޛW{k{,Y_o^Ƌ7ȼ@h 50$lOӋOXgc^45g{8WdmrAB?-LR}!)!)iHjE Q]_UuW# 0%As ۘI`1,Rg2DO71E4^JJ]L7Dd%0l\ v?8,gqݰö!`v)=.Z%/_Nǯ-h}Cޣ={2#Y8W;C+nJJAzRh w/\o_tFǔϹ,gVZtJ!&H?:ّ:wrJvRfL:M:eDӍ-s^8m ye4Pt'ؐrZb&>3J;f2Rʃp-1KȳO6.=M#T?6N=?`Z.mh>=ty`uaYV#W`|MM29PLqMxpD PA@K'AY[d tAWYS4Ϝ5x)TD[yfEw*5  Ay 8e>WKƝTunp 늰1ID$Fa!B(*hPD0^A8:G='7PiA2\;IH"8`b'8U1e=hT {6 .ρ?ޝ懷?^O={3Lٛ9{5XQ'M󒷪4nʙQ>0Wt$K[u Y$G+0EU~0ܬFˆGrw G N{t-wMۡkڜKۜrC_7!E—*Dxh /?z#:fKp*Ř1@hR<2 vgM"2 JIQl'_FT\*Cnb /]qq4 4e`hvRe 0Q|ha&cp09[nu '1bePQE:_)u ̲V K!zƌƁZ 1b"iiv*gPc OMZ~~0A}}~ LOl;w=Gbaur 45ic:uD9Ey:ew1٣gVi׀lo^}E\kܧpܱƶB @!K㜥Uj4!I߂}"PUg J_xQ~:Wy |L]8pp1GWF~G;֏-.G&-DuVR g _o^ϋg>(Q|݂>N0(dZ@T9{u25[}x_d~nd٭qeD6SFzR}T~Uiw'ἃB͟fߙ^I_Bٛ_~ʪ3W,Ļ+9'pcv3/ 6}LKլ兲K}fX^,[gOzF%* Axn)-DӥɝyRbQ֙ eiBa[T&[bx;+_Hj%bIMA+k杭=ǣwQJ;Ny2t.G eZ30{-#cԀW/$h{{M6R~-jϱ%8.ukx,1XF|0aLʩR ˽)7N/`Magjeh纹2v)}ٖM PjnU+h'GkQC#brK%bdF/cL)lE֞qkGZ(mz" 骇e>|f.٩'~Bn`2u@u_FQk %I/P!`V0/2ZDJGt`ZD"rR^VIXBg `*S:$O'{H !\F=rl4 beg$cZadPOGzHLdu4(&Q kZKZ/iNKbo i(UNu|'ƽBU8C'Tx+H 3I..F2ً8t`h|0ě @S,Et-\JjfEUXeVqT.i%3# >Nq͌A.Rg4ٲlQk 9\G_n{PM\1I@9J SkSȩhBK.#av@MT5إ5Xm~qnFn>օ*߾7)Tnb,>m[wv>I|{gܷ+@wF>Ozg.7Fo;m۾7>j}:ͪ*'KGWEGida>S~{>عyEg吞d"U=@lW :՗¾冟qל>]kxSHFUւJzRBB̠vÑ]tDzbxі*se'!jʒO$APӕ:O5ڲ1\dE'4'+3d>+x~֢ά g\ss5=e v5آ!rvL%IIDhzڹbaάם.3.>6LyD 3in4x Z}OT_ xD!|EYuC$8^%9&r6ȹQjEZ)"VD h$m~CjXp:@UCQJ?$Nq㞂OAbV!.<5a>~TUf.扄j{n qW}@̽vI(/DPc JR)WھzO,CKHi;F$㑅H>HԔ1тFQ4`pHnd-'S=uYi>]~`g:u8y-Xhm 4wpSۨQ1,S j@CԞ1gفZָ}Y|궚p*iML/m6_g/E}^ًqq>N8y'}q>NLu>GT4;4d۸d:r )T`hn)1yr %Jdc={߹G n;O>Ƿ1.@v,YydcOL( n%0lCbP9(? _ҕz SY~7t՝[_Cu5Y1f6dÑKl %L2NeN9)3oWE14Z?s ]-jܜ#:fhP&$҅d4K|4F ߃,;@ 7Bq kH -J,ɘ?ze;uQYuujnš!|ʟO#됙 ߮  .%뤖F rmJ?ιs7k}DiM552gBhorI"3jv/y@`٨o(}zX=FS Frl #eHDJd"*PC(]*HOcV` Q@Pq=|ӑs,WSkߒޚXƷ`n.BfKceEہp`OըjZ/YuX=x$7@[ׁZ,fX)k& F8As^[I^]>R: <@:^e#7ڻy_.}Ś~Ĵ?(ք{'"HjLC"}' 9 Ƙct|9-8GKy$"kY"Mw#>*F骊B/GPͻ0.P0v`H 1[,} *ۇ2_Å{ja]y[(`%T-Ԗ]n4=Gqcl3JqxG9hC1爴C0ǹgK;81r)ձ#B$&$yDQb Z%8i``ˁOcNP@.+ ɆmA6u.`8cQH̅54)r$L2ij% M~K5H:QFP`Q 'c"/)SG fpF' Ar`nk-t+a`$;[[imVG\IFA.(]'A ph9gλI}UR`6G?GwC.UKl6Ekd}C g BpJ15#=tr3/@^De[W^ (bmrvH4 6B>TEc~Vc秩F:(?D:$%է^ze{= 5Gf`Ȫ'KB +SQE+s*ߎףOzpErqY>x46P+e-UmԲERg5r'.;e֑uFjSq&|F.2 044Tߎ9/MNQ >j5ɶQʀGzyG ߧ3^A~(G"2ߔcuoMZt6ΊȌ/O|翽?c"_/@0p&lK܊Z3,z"s*jTˢK=BV20Qf#[}fݘhV[Xق9O B 4jc2FOO 󶍱< C^'t5 LpSRw hXqL!k)7PsX]zaEnB=C'ԯ;4_Rrg[8\= ?fnDd< J_&6<c}-`prp>vb/8S̖YQs87넍HգhJAKovJ;aN~{=.j '8#!y], S.IF5][#|7#"- `>%z~$C|gCt'vh㹲Fҗe5O8hQu<]7è,̞ ̦Ӆw UT p"#lW}!z hObjT (`sE?0̡,mX/Օ;EZx#c ;vTr0"DmO :5V<P@[z͗`N:;;o* PNy($fDK'mwPZ /_@æK@H:FhŨ=d7z޵9>oBO6b('`d'˞]~krMqЖ +80:lshN!3Տ2+P.IN4'?:R" =,P󝿡bW, Uc:wڿN>7[lro]!}qWzxUVvD<.? 5H9J ڎ=wu!Ժ9y^֞Vx }s~gv|:¹pzD2XOA4U1G*1kjE?K:mlmjBjB9yгeVZDu;iil6sVfuAǎ!HvsWKibq,&TJyT$Mtܪ|fE\{n-AVv_66wYShȄن~ʫ?q>3^Rxk}p z 3ҀcJ NVzϴpcN5 ^{.a(̭9 znHмwBE 2JpCGm iJ( rra8K9(+# ڤ2d/ (S5 `&&HAZXduB.8NeుǞ)₁F\`j#>0RjS#cA4Y3Z 6pZvSnk!ٷޮh~=1%anƻ" ΄n#z^ႵuS;c $F~ 2?;h'5gԉ2CF̣d|3/л=<~ogp~KU{OkL3v VuQ~\ȍݷʹ$NwG)&t;67ܹmm;SUE{n[U}T|T4 {I~#k.n'͕vިu۵z+ރIE<_={dU}Wr3E\3cX Ї_"ߙ QL)Y++pl;owuxA~o}Gu5Xz:OTk*%JCXcOS'LHZj 3^r ^"b>w7rgKH[~N/E5Ft.LM(ʢ4g=l(E귃DR2Jqkk1~Κ<&o'yyjdjM_ XkgMve;B !)X!\'wh3Nk`tsŌst4sn? ?1w1aF z5gI*Ir8N~vf:'`,,ߧG0Q) úza،[@f?9wS߇^˙\߼>q/V^PG:"#3untcn 8qܸ<:)lJC&(<6kr65mrWWw Q FmXADf Q܈OʖjneBўwoBj.tȃŒݵ@U#^7rVz=~˜bϦW4 q+GG.W{xs<yl@-m=B]H,_ܹy%a^?.f6^}<uSv@ )H̕[_b]+=͎W?ƅohC<*N,e1C4Ml= 1N<r "_*hL_ _L7A]ye68G&IyoɴFF@\<=9>fZTYU#RS9Ȧ éwY׎Pmj|=⩡$v@ @TZjR8nIV(BH7g7{wע jL|}x٪ T+`Z'P6%#4vZj28nYoeuBo% RpL&$+`5(NiGNu"! q2MN{xLA 6<@ZU >) I3 f^*%#N#8N{9IP~ <1.2d20 BD 0\8XCr.)N$#!$,MS,h oB XiF҉2b+yp]K|j cO5q`+ O(OH2R[2ȣJ IM'ӌgM=D">QOlj!7BD-|y7s I|j1a~$ ~/~\|%ཹ+Ekd} a *A#L]хbjFz>wWf4_˶P7* RB-D$BX\{<]*rY1?TD!MtsARb 0{Z}*%(=^L![0dFf%P$iv~kJD,ZSvT}~.Z^ /˦qZ\/Onm#P].졗:;q)Cm#dHWm?{WƑ@ldCv$/X' H iqC Iɱz(g(b䨻_UwWN(,@} +LTL|8z|3g7ro^rӬޕt8*t$MA|\j +)~a}zjюxˠdzޠpsݻ|6˳_雳/߾;Dw| ^@IuŦ74znn77Kq_*qHݘ3Wz&=~QoŬFˆ ɲ Xaj0ljSc[O ϼW0W&4]ʄ(oɉ GZ\/[chxG]1[S0ƂdA1ǜW_[%٦"2 ܲV K!zƌƁZ1b"i!i>n3qW% \Mz' `?_/o}XOɶxw$9.JF_.J(.J+oq8ګD}A.hM~ł(K]6ņf˟wD FYW"۩iAg$ XF|0ar {+%Rr B *W֢)-brK%bdF/cL-lDמsڵv2JKܳMį(M?'T6*S=D%$@tY$h) Ӂi]IՄ X ^ iXBg `*8S:$(OGI !DSk4 [Id: ~̚5-kZi kA)l=ىgZm%7mK,8ϭ{n][uyn][uyn][u`t%z*ST9i7*:2!p%3fVDB"WBh+H=H.HĉE--f!<\L9DŽRn&x92J#"( -J)'RHD#i"`^!vf~ l25=6S 3x 72[nvS]$I%J,9PS4YIEW^}|~), #z%;F$㑅H>HԔ1тFQ4`pHnlO5NQz:+W/ZN N=휫󥹚,$] K? FwriSNE 62( F刊`) NPJjZqAOqSuw_fHc1LK>tް;i0Y2\i7$P iNﱢzg-Ϣ8"6R* Pc>J SĘ>X)b4S@P. ;G)M}r]]nhT̺1?6+B32m`p:PR) 6Q`p p _x*|TmVYwʶh~V13&R/P}M,YR`$ _ G>(\ibQW'uyu|z>`Fq a7T j[.#G߾9leie[\ԺZ@Vf)lp` ˕y*j8yYGT'_L )O&ͲTv [M_j6?[ k"l) h4ىbjneCQۯ*Zԃ~cup6QoY!N/aw9]n'ys*EX|^K;|KͯN{9cLlr.`9VNF@.G{S2@JF/tJÃ&'Ҙ`Mmw;4"@n=5H15m=\/w?(UjU'M@*0Km*CA 키dsa>@t!oL3{R,hfS1.u& à+fR$yÁ _gREnR5Fj%n`SFxzkdխ]#V|ol qxhqAhKNw+VdطΈj.'Atèd?o⎯9-J-AƒE ,u-Iw^o/f- Wz鶕؍}1HyjܧhޭӭtTP^*N~Z^>ih$92j)Y0R`D$4ύDfJ;^w5&]@wXHǞ "k/%%\2N,s$Pt7:|3x 9͞AsTvԕ/O6mYpM)ɪD΁ٟmұ[aȭ5JǏsBܹ W$.?5q襣ɴK )$*< 8Nj΀/UoVn_;VJ }3d,/G%_0t hrxz[3p3U䈜+_-]f2=.I+Ѯ쪜OaǨ0 ,M֛G7Gu; =+Ȳnx|=(|2|[?ZZJWzL'8NV)1Y3hfV̬Ku zuJq:"˷{5v+ݵ;Ԧas^ଔl6aNOZvvV1Yxt~D{_~Ѿ[^#8?u;Z~-LvH >=g7hN{tFw:b:2d;Ǿgwf҂Nhdu_x$>vӭ&'P:yǡ=m(q@n*6iNGݪj {ޫ'5XvX dtA:MX:]_I3^;zyĤhVsbYYJ`xR|UeQWO_s\󙨬7iƽ·P8~W*/+uRL47^OG [;ʲp2-F5t9Б5c!-ܵO]]=һi u;]sמD[k*=AY`դOҿblwqf  uJ Bei@K,>K,>K,>[dt7 Kci!XH2gh"՞KxBC0LFDTj 6@M_ofC7,9tE*`MM nmBI5g19 =/z-iNT(-m *J]8Qj>vTV3*z>aB鐹J ֤3 K銹J%򱛫ʅlh5'2x}7^ۯ?9i0&z_ptAɕeNF]I'h/O4<0Km*eJ+%I1T bD 捙K؅D()G&(Cua5r& xun~0>,0uQyP 鄃Ȕ'JoPڋp!"52jM\XG(\\;')`"V "}ӞɫxWmM~Dm$_!fhuٞtv0=A$괵%(%q/)RIY)_a(Yw/&}?),F> ahn)1H4/Mz^_E?0|7nV fK/6﷨F thƣ߂jRvus5f0&b}]5\t[b8kisQm;.&GnO|l:!ؕ*}0j5tM}mH @À%0)bE3`;8|:;d6O83,YydcOL *`Vbai1g aHr2kH,M>vf \k.T>[˓x:wؗcԥgaRnϦ|@e䒙?/&W_N>N'?^YC0tm/.&Uob}&?j'T]٧ Evf>U!.LJ{fY£y6MDHm`|p,L>}2hVJ͹ (ѼI_TF!cyϚގaf @n_cbW[Ԉe6>o*d8-n_! 7(+\WQ[̊ѧݨH!4KCY ;^ I?n&)}I=E:6P`88̂?ɖf3qӯw/`NR+'Rwq{q6_'ǘ#񉖌CN[%FEȃEٴtL7_Ҟ,_N@q=&F^@ CPĄ "mx2,a$6r1|"b5R;TgR+VD!,gXÁa"S須CS/42d-qU^}6 +}j\jMmhFUOoaL:hI\y&x,.sKl¦WW4]SV8K?|1I=`b,i|"(h۬Hk%lmX6risG"`"RB-@!(#2`:ʿw[f#>6e\>)N#i-_nkË88{!^{#vγ f t3ͣ{ȶMlanYv@L+@WpyS0uҫDuH[m'V.^m<..j \q-vitaI%-_-of)-`tR|QKk{*JsVZ#liIb2.v)-S ~4:dJ^ĻLG]*.xU5a&чv2h_վ'x݂gT98PU ;#ޞ'=FQ{i59Ba"0nZm\xfh2zng:vh'zďSkltS/_^svsBSݕc;~ ]hk7Nkd9F{K ŠPwᢠ "Qп=>p-w@dŌ{˹t5+t;8ЩzeYE)Fx"6R* Pc>J SĘ>X)b4(`JXx:7L0R'wwc|sB83ŚB-FwAwi ui݃Mn ĸ8XJ/X 0rEs21/ʲ69 Ƙc|v1cgmI 9Zr;Afp.N`Ka/#`O7W y P0v H |sVLGSJx->c21c_枮Ŕ)vn>DVy|(rPkSI>&F7Rӡh47OMAy,7ʔzf]f]+9Cz4~7-֝xλ@&;Nׇ{mxiҮH>2(Y~փ%7ihPfsk,HbE*]Ɠc%OC9OE=qTQ&cDa,HriZ 8Lfцzp e?twàӵtea5>$ hgOq蘬Ҝ{rX}bDJ8'=mc}fCᓂWɌVYǼ[2`ZEJs^8m ye4Pt>E$\؄ AFҎYuZb"Gg={Ξ ?H4A5>[]}6}&>ݺ[k٬tYEs#M=@A ZT\|BD}Òzc N˯9PI`5\ /@n1Z۬Q^G*5)9IEdQTV+!$)RࣘMD(@ܲI}wXA҉06Bx A#zZ' _\N܀%!$õkx(\`,BG뢢i0@VQ9Aj ~ 6 70|{>lj_Ep}3"UMh3}j!k.9G4̳$f8IǸtQY>F ŸNU^\uwJgk^_#mzٖ1AErD#@qDRfD.ѩbLQs;<*ϏX{ aYiC(XK KL_@FrC¢0)KRa,73M第1\fd$K(d_ֿONyB,.14%?r{!~m,m~]8fVTUnK6ִlUͳOf ?Kw՝.g ֕U3se4Y[W]sM~\Wgfw&/B]#q9#w Cc,o}LKLPL|:_ր.VcG]ӸsTvZ/kԮ2ތZ${i$ }:>G8Vu>t*N?FvFgo?|Oo׿~D}xW-H?u+--k^2Jg?.|;ߍ \ǣ3SdYbg`4Lt4K3&Y]0΀y\B2C˾ꡩb{ Q^|qiK& ם(!nt]}Ornc촑@Ζԩ|cPbe2rmе1,YT`#)E{g_`6$` 8A]TtY"Fkl+9hb*pn0;Rs¼$$`hiF,z4WR0HgDa iU3PSAbS |#('@ֆ{@ Ҙ>(슋eƞix BBb%#1)DXҠu0I̥>hi9P@iҞ4(ROzG/0dQ{x.G]r{lzׇL}^t'BC*_NzEE-5y').b\1Ogw0OS oz1 w.7lXPnV b6)*xd~zUuUJ,ˠAF.et1kz1~cK/~'Iq8 ! (%NbfsP@K2T4=|ձ1C̽ ^))89|F8>썊΃N8x8͍p6"T!1$n9nx;hF@E==' lJtO}`JZv|_k?ǗTƠu|˛w$*o]_?}$S`=r$R# 9>Csi5.M\~}q%,7F,ٻ6$Ugwu`\aB?%)RP}ER)1nVU*vwH# be}0z)#"b1h#2&"YR8WM=)ޛ`W󹹞^TmN~MwiSiaE 62( F刊`) NPBJs䌞2b[1ٻeW퐼VL\bfZ3r7c &ĝ4D̙P.7ڛ\Ȍg( "#b4ĎњHKmUh7j1w*JJ$D&b5aE1fV J2 6vZfo3`+ԩ/0ƒWE:PB32`p:PR G^(0lp؉TUxB=\Џ: u6jH")3Qu69L0˧a8^T&.$Qs-9˝R 3i ^*S{L#7FL?[a# .2<閭G< ʖ|ԡ c1F1sɭ`f$Ť::_sYo<:2``%uSBjG)cRGMQh8jL@" QG:u <39]7 `3f܋ߴu?<}%IZS~9 {3_.5,xi6n=xY ^^WIΆ#~/ sZD)3ln%9XeG,υXXut1RYfs,/b TcoCZ0*ff%~f F9JmRgn ug'9zÚVk]`\mj x,/.}jc՘0%OQ^D.'ǒ^ Ĥ0bA BYF%,Z"b5RNXH>e)dhp8%b"XZYh6t>sЉ|G6ݴnͨDersr7CaodY@oN3/O.H0>Oݭ-h];EX=8vz| ,w/+$^0O0,^B[';P/xyqcjamw.n"A[jװy`7ؤ4ڞȎv.t\.E]jr\སۭrU.ni$mkXb4z?P|ޡhS6ik'{&ɍJ- Tڮ|齡3h޺4RTl=횁-T۬]}튾ª`6U3zʑ?q؟v87 dɽcq $A "YE )Ĉ;6O"X,3)Ybe?.Ct -Y )$~u :^"ʺOݰdbPW1Q"gtLr)OZ󔪟!&H_t$#B)`?AZWQljP:M:eD*tu ǰ!0O&j\r06!`BCbQc(,<`kY3r ݝt(ֲ}V~x7ס?D m>uc|,auhΑD \* ~BJ>h7p$(˝|$xnx]SK Rհ;"E Rr fTDHfEPJb,A"A$pF <*%C㞇Qά?|º ǐ 9`l0"I :+ES3kӞxԳq;r=?ִ X' Dp"x.*@fSƈ R u+Hx!; av~="r}D`+W Xs<"aIB $s]TָϩQʸSdriޔ9"ȟ/yUQ>eKd}Nc#\s)(ԝvx{RLA ^wwaYiC(XK +K6#‚ Y3,qQMOyl|FKVf@n+S*z`e5tzy?O%fq'nV0do`ͦ6 \^q[HAs-̩nUͳ3_Kw՛ﯧl1@siFq88֖껵s~_lqa!6#1z˶aH0 3,A}K+V1Yn`nrJQI6W2gIs`'.u}]YrXÁy=6T5 :י\];?<~L_?~ϟ?aî'Xip)vۼ*--)rf~~0(\:?mN SdY\0ge=sPduaģ> Gpo|] 5CSŶ.g]/6.2ߖMh6 Q!vm~Ym+s=oc2fKp*ǘ1@iR`F߭@y v%AF=iR;_@: NDZRJ(V$c^H4V%=J +z#&KQvbhLuF,z4WQc9Wʆdbj$xoH &s>bVr恼ikE:uD9Ey8dw1٣iWG:1ఀ1T)b+t@HmOjRsQ b= c^s EsBmV!bUsy=קշj62 KU|.jdh.sLo}ZK&kfa8Ԇ&gDz:ЌM$7]~Yb"-S,k5뫓k|8XnHRVa S띱Vc&ye4zl5BjJ?@݃x'$_j8:A{l4iɥ!r#RL@+Gm7 ,3A.vƳTMlIU,6ůf8iJ@ (L`xN0+D{##a:0- 9K{]IXBg 6lw4H&n/@׆t9#&Ҙ.$C^^i@ <]Bb%#1 )DXҠ8KFbR0l2b%^)iG-a{ 5>%*;:IquxQ@lqB t,uzVIا'>-PUIM85yWxc<%\'"o= #W'\~qԪ=/nsb8*ܕ0.)eiʲ +7rf!<~:!]}~ ʓcx6 XnAt輽mW<[txw([EY6zTrͪo~+`3@A5o&#L'ҍoa?eUݟ*dAev9 l_<=ɀdFd|3H)jdoobОCHM8ArK@>F b+څyfHhEȀ.&q}E֦gjW;,]ռW`twny|׸RMs-d<^_vsYؚij|Cހ=s^lƚf?k/m= s_-98уM˳zOjәE禳bK2S./+Nިk5Eo>#BDEjdԚLbBQ >j%" em ^p휤Z`;5)PʍK>x92J#"( -J) :ƮvFf4tɈOIxg ,~j '֣?`}#kвw;@m\2".=M/b*u9#D0R# 9ޣpG޳6WPue!MRgR皧3ErIJ6~=HB J Mbth$Ȣ: Go[‰NQ% FZEdU0bȽc)$)V2""&ZH0<)c"-[pelGۗp?9{],65m78󹹚^cj4>o&n`]EbAzQs125q9A<aRQܵ{MiS iE 62( F% KIw>:dZ+|%k2vՅ7We_9~+?(cg@ÁI|:|?R=+99K9j8O@rm˃w4 oW/egY` t^׽qBIzp zKM1O*Arђ砤PJ' ho:kf}U`տK(< 9ȭ-FU$'_)t l L4~V8+>?.#N0 4W mI[C'48FvjI66:;>V1&Rwo3O 㢑kmы^x)c8CMRϾZKݥEB#K CcU9W50|;W4bBj16SlcfwzYI 9Zr;AfxfuqՒsP1xԏ5{ɜk&+vӍI[|n)5꿰ۆ.Cqؖ.[E6i oii5Fobf6qk;;Lvs/6RF"?L Qʸp8͝9qc0o67LN_-Uff^ .VA/xZ]xjpV+;>8l`͆XtzZPg?"[߁l.hú#`DGp%!A0MNޅѪ8o:Y4 646 yBCY>l4DߔѠɢgJХ[ޤH@kަ 8Bߦa Wp';`H*Qߵdi-"0|ْEX斳]d4^@Mf* gC!-.HmI!msmvX}W.GJhz(gXqSgfwvf6$/ȃY=h/cRCA[l[&>i^ vJG cB&A Kٱ$$hOHP %p =*;WXĩ>!۪-g4ɽMPJ9ߢ3ʁ@'=;rB荦+~9EM$N(Eas R"mC Š6MT=ՠjϲ.zIYy +3>lCbOm=+xti`fպfVp @sE,oP3]ZХPLI7 }R3UvשVqf붾ܛ/͠ (Lh+*JluszbRLQ/ U|+GcWY7&~S#3Lw6tl c<5SY04\`D$/h;'0˄ne1 B6A(˨47"FX)uHHJn ,Rg2DO9L.DpCLDK9 Q ƵH5p7kdKNkr:D`<u'Ue7ջ{|ǻ(DrXR|$A "YE )Ĉ[IgO"xaz53}[uQh]n& {cխgh6={Y_mUŠJb7M_R.iR9"TK.@FU r:a^T7q5 2 r&ˍ&$2AKg<;`^:@D[KT$Th7Zs'c'Kib%2{s|.R1}RhR{k%Mjll=X$S-˹tɆ`\mɥ1-_̢^Ai/ Y̰RD M>P%P##€l&vrq.}g5bxzYsZcq:sB]i6{{_Mv p 5BKZ ?@or _P0v`x Ol XHez#>fL[1o"gNe(jF[p hƵ[,gGJ؄ A % Qcme)AɵDX/5^#Z6p[azÜ_\birP2=]2|`^YhglzW|lRQE=Љ@A %ׄBF}&Po`[N˷$0y&͒@3ב HitL*" VafXTqXG$)[9YP~, [T|e&C1q#tX{ L1WάN{QGOBہTSF2\;IDp"x.*B4P-WFcDe]2Z;΄י6g < "٫^ߗ"GR-ED>i6):"$I]b`ָ/iPbȸ,`&ٿ?\fOSxoLjXEcL$Kf'X3S`)"'#tY J1u'f2 {2< }Wf*keE-6`M@ZX.*W#,\ OU S0&1eh|ᦈԝ~ `٧z`tz9aqk4 iiNFoW+.@)UP;kFgfE'?~^T_׍j̥~E[^M 3#?0ڙ_=iiZ;MX#02 04pп!zwٛnZkg%hEuZ:IA041إW_M@~,EO Ά)@]_* zqz ?ݧ7?&OO?y O#h?Du QZxq$B!cKne?!UbL4Iɖ߯zHCRTMneNznn'Ώwcfeh \&o;~xۆ;i#^<0}X {On[\nM|r_[Mh~_wkCօ<λz1#*fKpa(d g;30~`O?Ef gexPQC{Njvm3`o uu:%(V$c^H4VwClb%t SACt%;V93Och)"RF #HFΌE\S+0Up/JM޿ >ra(яv0zx ?*xXq|I+ayjP^Ԓ DAt jg\g~s޼kNfz>5ԖisCmw |Xp1 cmu~j>vݝw+? f/jv3/w&w1~$i5fmWèy૷GnZ:}_%ӎ0˝ʛf*-c^q(*`+weU$XLjA$ב:.759#]B t.PHx3jg>NLj9+4$ ҕ3u}Qý>7?&X"f^-]?4-&KdzTJ JπA$GZ.tDV"DU*lw1YݫwixT)@oW2Jqj]XqEg[։V:aWV˛^Հq5a} 4hWApł╉ZVFDh_P!s9NЧ۟|D?navF` Œ@(]H̭3<7>~PZ(τKLHOh .05tZry$ )[X1c2b=6M l! >Maa}%e r8D%&5jajŝ Q cHjR9N`.L!jeXXHEr]E\(U3zo=#nɥ'y-0''v&YϦv eǩvfVye{Q%&!s 01 h1B$eJF(JH1rK5bdF/F*e&v˸fB]lE¹D&wiA X+oO?ݴoT|uǓW`sShd a (%P, JJ!#a:0- 9rd+-;0Q \HA803jP S E " c}94&=Y,:VtV%dOIQHLd$&E:łKs H\:pCZ6M+vHMh ;R3s$&jzgpLV`;eTTD9s0ir0Cg8 *WKj r,h2LJT\2HJ̠ԙ$pJX"w ʺrC'YGA^/UPqigBS7޽|QȍCTCS-]L ZWк~ԺjZo?Z ͮn||]N[hK9|KބۆT!vjͧgMwcc7%slWב[Owr RIO7$G^7*:$p3Vu0DBwZsQqDHZD ,V!(%V" 'E:&<ج9JqTQ&ZEw9lVJHH!$s|M9XjBS.0JjrWڱ(/2%L&))(DJ3Pw.f+_|> ?9j *f= BRa*+-$FNE+rϲ=TOa?-"๯T]?l5Mؾ`nWiT|Lt?.u"<5'i#CPd D^(]Db1OTF1ZleF͝QFGA i}tdll8q!mHk_esb٭h+JdyȒKqmHR|)\IZNΝk#I)X 6 uMCL0_.Q>v7@./V lu"[(~wnfgl>ϝ3T} /+XMydn/]> xZ~~|<$ ].Ӛ_yvhp?Uo.MȟH_$B8/P-8K@gk(P놛 wယ2[:%&0 0uߥ\_mp9*Cd Td__k7V ? Y8z(wbhB*&I9.+6_馿:L|_zUop}⅕MZ͞| 5G;c76A>BNP~_I:>mI6Qw8ۼ45ЭGL'mͯ{6f],ز7EoRod =!"}bahpP%J7bi1g;Ja8e|{QU\s|?nv\VN.Q◛`C[_J/^^h,YtY mѫ+VQ#+Pщ'sy1L]W\6Rs.zw!3y6..”Xsr+KTpDM=A?wSA_P`~aXXB_z{Ь?r=ll̡LMXiMYq˖>TӜwmn;17D5{왇 7 .=wBgȁN }MGnCc[cQY@yU(j\vN2~zy|DrXcܡ0$*I HwVQ#4>zJ,61bb4898Lp0 /Zqya۪)$GK&ډ2јeh(h;?G`Vp4AiNkdTL(WM%Id3BK 9紀=;`ZHDϥ 7ZRb RDU"QHU"ƬJAGIAR kz6e/H6f >VX*"f1JY%\6Q`p0+af:1"H(`<<ۜ+u9#y6W-8w{lC%=Q+0xגZ^ ޵$-Vλ$}kJ()|>xTȅsc̱j2rT^r쬭" bJK*1r2 [cR;ZH 欆)Oc.*5X@UgI¬S./ޭ GOw=_:A§&;h'>Ng0ȵ$my7?-Vkٕy35܊?wZ$qM]\iO* |m+\it0'96JU "e*+ bRKkUR0.h& (Eblqp4jb2TGQ#g0:bu`P~` 2mUunϬVXj?lXj+HZZ :: :҈I>`\¨b>9-*MEXme%Cp2] ѩ ѳ^V.!xKXCFN_֐[CZh[DZ)*sgC{934 .|I'[G+Mǥz/ +&[mN3>C}Dq=|(D @KZiu?f!rF!(bRb*[XVNf1(7EL BІ!4u0Z~"b5R%l(=ݳ q~ОX@fL>|#21͋#,gXÁa"S須ā{ꅕ[Ŵ(o坢W >}z%y<|`{Tbߜ(\⳩i&)5Ӿa^U ?{۶_!K[P@R7=)u@ث-D|L9g%Y6eɦ-ɡI.gfgfŽ,#~^ h|D܃Z"Ey5Ne(jQ:6<2qKFE:UMP1jveha8zH́|~\4u +f#~8ԇrl>k@̍.ρ/ߥߧowCL+0g%4Fn.Q)xm.4*|: XfH~zS7__ԥFefC.xE@!-@ˮ@4Ulu9قov9MQCl;12K@V;GJQ&?dr3`oNA@*vV} ɘW4FI6v *+z#$+V{=${^qʧC2JF}0pf,⚂\XAt*=9BurWe:;/裨ٹ2y:`ox+z|`6&<՗6`e,,:1d f1g[G‰uXTX:޵f/w:CHc6AUyå O=cF@1gz4`cih}Ә[)kӾ'5.ַ 균K1ٲ-oގT=:z ᥦi^=+cO*_K=v6{ $3J,\OG:zYi o'3ER,T-+ H'Ma;Lcwpwlœȱs@uJ* Oo^&u2IDٟi61 0RT :j#=c(`aZ h AId_qw C}šP_q8W+C/t,V"[7 D P{d7_#.z f4TF3IUypo!(&"0KYݹ4kjzzWà X_}iC݂f>#F"`"RSFJ ZOQ4`pH~1Mm)+k]Dtyl]b֍G\~磦=T{P"jC%p>r^tkVYe=||l>m3*wZ5Ebz_Jfe`Zafn>2/_/+{8g̞1wǘ+uϡ-|uc⅙bmE'*Z1WNJBj-kydFL]ȸӳ7r>i/Iʭ{Iw״r]}5˛r.(͞ɯaEgizζ#5Ͷ+Ƽʹ!˫їݿg|f\1hĴ,쩇{h`ߌyg#&UVCqlnjji[UiiNHtc尳:hƳI(F*cގ'FDrXR é!TIEBSbӇwΞ(Ecg⎷#^iM{/vXlrzb)}BDvja9Ȩ?/*}Uewq5 2 r&ˍ&$2A -"6R* Pc>J SĘ>X)b4~L(ة&l;Fg$S_];vM.=EZzhMA[0ВT8pEb&JæT Lj 9`π0^ߞƽ];2RC'lȜ#WΫ7]f-8,ru,^]{^Q$$Qs-9˝R 3{~W:nY&r[cR;PH Y JRz\Tknl_wJw]])mV[iRHbHQl*܁z!VZ8vKQ c1F1sɭ`f$Ťץ%uSBjFix-FMQh8jL@h#v:M1:$WR_!o}k+4U`}+D*ILnv#sB^@;]V%܌a]U?@%ȉR4ճ5bC^%ڕJeNׅLD9DŽRn&x9IΙX;(Z)"V?"D4އD*iugZI ݗ0lEթ%3,4.3HϤi/|Sa᪪γR&[mb>Gw;WzqGJ$Tj3DeZBCib[%,9[})dBmfL{i?v sOYVr(- >t@BUzT?.B1*7QsQZHdu ǰ!0O&j\͐r @JX8FҎY | \KLR5C0N.x(a2r(|9rn ϮD2ZOIoy-{];r~dl@N R( &<8X L6zc N9PIÅ܍dPnH^G*#R"%1pfEwZI%pD0AP8ԯӹm"ԑ<vח ~u!GpLh$2;tVE +gZ='yS_jkڄ NR J"8`be+drחɑX1|`)"K48RLeA9!x1c5keE-6`M@ZX.*܋FX9d䆳`aPLO01>dEZ,}7٥l`e4DX\bij%3p= WldTM*sBpJU֜fOgT>V;o '=j̅6*xtv^-Ջ]rUɹ);P$.[!1~?60B`чbR(Xi'fhnrB%h=fmPʀGj"C# 0ur%ȯ#<٤>\cA%^3O=بogntyNN~y.>};?߾;b'? O߿]Xip)m^r!J+KcBQR&:EL q(톉 +0Xi?*37uaē͗. ' pN>nZv5bƨY|)unҀp݅Dq@_xEi96ZiW)^:[RICIT&aJ4.`ljV;GJQ&?dr3`oNAEeN,X+1h$+wmb%TLVFK{GIW>zH F+%O"e0"d`X5$ T{.s?jtF_GQ_feR6u>e/v@vy,(ehdxKjY,bY)]91[]nh+酔nDeM.oVoʖ3f?lGqk;^B?y$ )[X1c2bػ6#Us #!8vs%E6YS̡앃W=R$ŗKn%NW0MVHP9zm^x?{<>7҇Ya5CNyGޠC61զjSNa(y1$5IfXPD|0aRx& T)J],9E4i} IF6>l .XqT>Eo} Q^lSn'R#V"0c)4ƞFc)aI$ FH\;'3LGL12ڨ,Eɪh-J}P=)QÝM*Ffd Orkll֌QIta6U /JN)J6zLM=oXvmm?P^o׭Go@cghd@a2 %P4 J`!#a:0- 9rd Ζ6xK(E.qVqO崃a)0E^>rXPAC^/{C2`I!@SAS/IN q$53€r؊"iEҊR.Z |ChUٜ?N8a(pu権"*Ü(-ag!%U`5T3[Z%#>R^ΡzfƆY%-.^ExH QKigfPpkLRX8:D,4`A'&U oVvSqi3gB6av~|ȵ14~ՃI@hi֏2 ]+<]G_ot:fCͮ]̈́0Ӯ[3?j%" B ċQ <ج'!k}br649ݛ$a]vOyc|y")"qwŴR _HT!D 9P?K Ng ?88("Q{mK8Q)7Ai@.ldY'aR/RuxLjd<)V2""&ZH0<)c"ҙg6r Z㪇vasKnjV7M/[3y+89?&O+Luʡ'vZ#8jN4>FF1N='ȀA LQ*" %[<^6Z1 M̨3hQ9yR"g9N8qU鵋. dM1 YmWD=xb~Cd^J[pOo |NĶG ?x|L HC޿>Կ>FX.|Wxçf<K*8 ~؏ZIK$a?*LhS\4 Dt 6FΚt_fl&>H эpmCmcLJpƋ<]&!p9tF7n:??j}s*O~^ɕ`&W;Za8c{7|O~w{$o.{7v)?wSq+6X;EǮT^ex1 rV\a\1PeNH.ExvxOĺe2 ./y *S 1waEĖPƽuH ÌМ9*K|y"צ=]yt3|~Xvc[m.>MpvQ|J% JM89+@-=47o<'*E1E9UxAӫ'wSߥ??7-sh57\w R7[Y>fp0 FVhZa0 hTRRrڑNQ ;2CTL{Q.z,(DXZQDHS\dʴW:G$}V"#nui(M k,DfV uf?߿JYj.jy"HCrEX#Oa%"&9D>K@gfWu e%;Н好t^\ 8߬3{;;N+v N bO5h`Ta2-GpqMq^)$ZKFI. %A$13BLh }1'щ\v1a܉Z$:Q)I7y53(rgPRϠ3( J=RϠ3( J=RϠ8v,9DT\Y98/ၱXiض*lPqx )&Űܧ,ģ]g)v!Fc5Q4yPcr& < MGb819[-(7b>m.4-;roFi+l0v]ٺˏ}̆7E7ǂNWuF}`]°b>&aN(E+fB+LHt*H|r-r<%9X0Wi]SEF;^ElQI 0lHcFOΠwXuB3zj+JcS ;ɕvu̜-A5pݑiF{">x"H5!sͽ륟V_? 1D(U`X0J`D$4>i=j&P 1aAYˆHkmcxJRz԰d =!"}jX{))qJ7Dd%0l\llN,hWOZ].TjtQ >0U/9@&"Xb D,gXÁa"S須{ꅕ[ŴomOx DE˪Cs%7'o<9!j&~;_ 8z(w2yeX$*< 3h:s=S4zү6)ąF^|=3Nz۩ӷ>2kE>Q߼O7.NOkC̱~2n}Tn? XΑXٿ]?AVfF)rsrs0kE@f&,6s2 4тi5+(U{x\C.]$sd,y:\ V# ųc ^ߘVWm'mzՖ6wfa`ZWi͡L^N:-wNZn^42m֭ ^qGRRIS= EV>0MgqnGk(jj>s0_os~tk+$ځ&)ux~5 ԎӺ&1~;mgLQ)MTLǞyZsvްռ*l։褠]l8zƳM@0U o&I7 wGB$Z[U@(: K"), zg52H㣧b#&g gO"XagPJ,(1ˍerop|.):[u։~f{D 0 Y̰RD M6}JLG8F)""g;y@qNtW 6qn;5qX4>"T냕]sۇ%zZl`X`FPq.`9Ve;kHEҒ)E 0S:Uvn&9 %Ǽa=|3<%cD)h]A>;^=U5%ޡ 5/jCemtڭ@A*&grA &zk&ǵ3:&[qOY~k$RZF{E:&:9p3cRTǼ2"ʀQL*R"d1lcS,  3QRl6!`BC ǨU1kQRS%&zqy-S>sDIUʏqӪNI^/oʊӥSOzWڏNܦx2PE=@A %ׄ? AQ kԎpXAgx>\N=&G8ayHFǤ"j03,*͸Jb,A"A$pF < ^+ Cކ~Gmkvu! GpLh$2;tVE +gZ=Iؿ~ښ$õkxI\`,BG뢢- a&@<|ʕQ9AR^A6;sη/XgBg`z\ W@AMMqHXs)<"a|ib.C qӠ$ q7Sw1z2"ȟoyeQ~,ULv9F"rnSwe:U=%xzjMW85iatH\dSr,H}h~H\D>e`|/cxw?]^g+;ݯ Y}VfY~A?ht`ZŴBwGn!Wuճ2>kDFҍ9.MXW3TOxn|wgJ{8_!K`G2,,^CS"%[[=P(ixIc:Q_:tݯ篟ޞ_N޾tB>y˻OޢO`}jڗsWϽï:rۀqJ?kG ϧgC4|]&ܙ_>>bj8_=5װԔف3/os 8pLY)h`\AYA|+%P{]Br WA?*#ڗ6=dx[!2 aGP_n>LiJsx'u ,"xM(lI%nR^S VC"eX"E4eb\ T:RDCZ[,OsRWȺu_bFZCo10Ά),ZVwaDT]8M6"eM M=kRr~< -8hXb $00KL,2KڱHV%*#9Bj EOkB28AV 6j[2%jYeP`u/'/@b!L p4Ȕ9Yh\*/IZbDD02W lrX'^&jÒz4P@󨐙BNADs1.$G **k;ˎ:{Aݗ>][3353-9nriUfmy/ |~>zxʩ)YgFpJ]IY-7ArafIԷ;^@z F0Bo5rb'/Yh68|=o ԃ5dXpQ9vhGHH*Qbɧ,%  Y ((aY0yY 9SLvL'[B92΅HenP I'LIExb]gVhZ&ݵ U8 I/Ӥ7ߨ\ aO0$zn*?U^ykGs[yEf^D>;ر܎;F6o}(Z\ڨYp4 M. Z^8l!)a Ft5:ڎ=j62 $2N@(c4VD&4ANZS9fsO✚V=jTA҇*&hUtA3CMP3@p{}DU s,PY֖G ext_'mxpK9*5]Tay>vǝϷ dJ^d MO4u<:y5y\sw vSy5,>@B' eB\ξٍ} #򐲏pdn5YGy{Q^첏^cQdtT s^{Mˤ]{=ȳVGYet||1cbQw 7͹lS*5R!Gy~JWfl6gÙcv;\;m3V[fPwhbj uWj'҇(6lXY-i6d\Cw7fUm:)ORnIUY~bsouY-th 23x@9O\j; bܿ:s4Uj8xQn>>cو9VɬYV{fzGNPiAYEN:T%;$:^xn"h;ʋAQew9mq҂o~gJc\:<OI#JK;ͭʆ8sԦDoZɾZqry/r[C.l{YC,"t7 QfF֟!Y\(F4w乑DxcxTAVԾ&B6IQť7s\׷t\ߗO\KsD["O'!Q\\ !)*pLNɔ,(qHd$*Xꎗ,S&)|.]6Ru*3/kZ裂4`NtTip&o$)*}JLA v  M<Dvm->{ۤa4܏+s uwFn.C1PS!/WC)ǐWֵ/ǀr XuзC}_kc%B <Š2 &(AsEbI(Px0+*^>8 U3p CXzחl{כ6SIjX&*셯M] j~P[PT# PC/8^Sj8m=&e[s"٥$KwlrFGkd juM<ymeRd1:q&Du)Fd#ַ[4H9)cƒଶq.D`*L M-_w5tnJ@Պ'?M ׏m[]6|,_NwIZ!ia !b!y e<,j'uCH/:Pe[, J&y )R'J3U5a,$m@x 9"Zd(<ȁ{cEPHqyY!sp"ZJ!S"~IQ3 ^8 ,wBwԾ-մ#YaވD *$"94ύy䞂D P¥莑er2!/Xe bo?ÿF#\#%=w)8鍾V1qDI1|`/ 7S3_w].;)!jyY0y8Ͻ%c2|7GT1SQB9aGTI)G,\Gq`{wzRVQ Fb5*/7\ɬ]I8S*?O珑 1loV/wa+A2;%<ٟۏys<9ՑȄ%s#S8e=2ohͷ6\8C'7PP{4{7}<]o/'g-{ŜoaOϦg.8Ԇ/'ݙw?!M3MC< 37MÚѤqEbpT(<|Y˛9ɪ8+#g74k^yDg-@y$O,|~}Dz#~Lho43XcA]1D`_!9N޿ۓN('~yw[:? LWQMR74~.qa_z]Un|p#\gmhz4_lƃ>?? ;SGLښ竧֘6w=;umnaߖh>?ш!>F hcvu2*F)d!P a}XџWo{'Mc?&p "<hͅWܱۖgF6$$h x.jLh(ybjԻV@6qJ DJ1Ψ۶wEC97 Z΀) #p+H"+(oDJ{oKV.*58'_E|Ҧy71`ox+#V aGP_xN}R|O_? @zg1i ~<D#7Q2@ԢɨǤ 5s47@j9(:0\T[<Hƈ+ތ8|`;XSfn{grn|NoS16"1M jq@ :g(@,J@mrpk w TN1w ^ ('.+)bB-np-@Pp@aJTd|Is|-h~e @1.@)E/^L*?ry_M[^l:p:uUܢǷGWa9^vqX/??O2XϭML(Ut,1A% &QU%X`$QBXQ(J!$HtHC┡^{-THƠSbTr6*m2Rb cmlR|o/vaw}⧃I.h`0:xLr(FA4iFt JUQRRC76&p",<NC5;8fK%8 A1TBGI L#x [bn4 +/V6jWJB*FQ:1KbhE IQDECRUǪK]T,eEry{OTG@q(|]B:oUMvNMi ;R,Iu?&꠆?b+m7p{0ئ1R󙯓L)QP+ B,.׷R蚥YJj,tTJ3V>2I$^zN5OmU!Ij֚D #g\Ir\2N*<%`I2m%Z:$=9[ߕyUaoztbv 잡C "< Smfܶ{eHu,CW7xu{fvq:=.Cvf; ^x#?_9X9`+whcŧ=!k5g`v$~_]]\_\zhO55oNwo%al[ %^)(86ofF3}66?.xK}g%r8AK߀K!ļY:yI?O*2tE C2Z&%Hƌ,\"uUq.@N%eL*$ <#:j4z9"<҅D甆n,tNԑJjfjGuK\q~ ~g~v.םb^dʮP=fhD4"(R+`ven5$й{3rL fK .HlrQxċkeE=E"8Mԉ &4n;&^s-E̓*!:E AJS ĩv!ă2L,h5T"=&b b}9v7|3d`@i.noٟm|h욕LKñ|9!?zp {F̀ )[ Z'T he" _q7T$4 h(Xj#jk%NR^.$Ola .&~ gz0SGկ+o;%vN S@XX eq/+Ci5}e)9@ 52WY`Eo\eq5sҚs2nUs͕dCĹy-!ϋ/7/Jh035!Oܭ~Qp I4$G"՘,GR'& ;k`ΧS%O?F$V ژ?6ϙ~-?vj퓧dM7-O}qߘŠodô XgkHŸNf6e1ӥ牄(?ǭj 6wU"i%%X"$yYꅎI)j'r4\j|ч?27!xaŢ) L3M\ΝT2FDKH5GD`;e 5||y`=ry>Yz5%nc3K=[YV5 %H+ h܊=e07uڽgWmᗇq/к. F9r2n@WtVCl;ǿ[k nkq8:hEemee;r Pg`9y0r}gϭ#Heevn5><%55ro, ŕ2WP\+Cqe( ŕ2WP\- ŕ2WP\=^Só _Y2WǑqun xlj+m哮|UӪU>']k"id20֌"VJZ Z+Ak%h^ #4P Z+HVJZ Z+Ak%h*7ԅ n g[™ov|]8W']u哮ߕOIW>']+".W~~VJZ Z+Ak%hV֊,FJ(.Fo4B;90!>LS eEHm qQIxwQq*wɍyr|%KjVur%PΖQ#\hܜĹW2).8͚ӻwN1?p39o;g y<{w:}sGrWi( ?ӸPo}yJP+&AyS73HJWшJX#jW,8颌X%BhZY\ZPI=sjFԨĥTL$АA3CMP3@p&e6Gb% QC8, VPOLO͢!Ӱyz7#3a˥"8s d *Mo451NDͼ`%t" qEw$QݴޡKdUcp)TŮƷ\ 0M Py6 LW 4^݊o<;Kv{Zgl gx92\n>(h sބU>55XtѽQ'fuG}A3}LV^hAh<wC$X(\%@u{q{ uT>3xpqq"ǏmQ";,/vzDO{H{GYeUoRo~M *Ȧ߼s0MCnhj&%ʽ_p@wKeWܢU1b^O17uzeŪJqax Ux7%}Ub[>^v~b Yh-aN/z9~0gVOVn#^-qWoƙ əeA|X=C~uaf3bl<_xfܝ Ɏ :`xM8Ly#\lyQ8d] SUZ٬%}}f2xiW(7EMP@l > >t)qHH <+&Vh*-I4**RQeV+'DqN+EZ c tzoz^t)׵2JA Mf:Q395o 1h7( )jҝyt'~|gKGx_HTGҽwdHM4-'TxIuTgBtJd Un@KTMMQҺ5jW/Qze©ܘ*M0;٤c$MW%L%Q^Abk(NKBx6ϯWޒ8,7H @lƒ0)~e9tf22Bt8(Ql;$Y(ךY [Oj; CvT[l-ax=ݎn _Qxc]qʹ='=G˪uprAxf6jVZ]+THn WY\*nUwR-vR¬S)X&m_ ykR84@prN&J8l7rFN@O`j75o,MO2HI28\mr遺#NUq6'y.tέ6FA{p;m<8v(Gg{Ϥ  P87.Ow>KZqg0{c?˯cye|C] ׉$QIpi][oɱ+B^<ܗq68@Y}S̡l˿T/( ICr(g]nLT")b.輓.y2^CHu]UZ ԅKcqm*UZKhu݄0pьipF;AK>Pxp(Yw99`L˅srKu`J'J$(z6H_ݏBڏu5HNڠFH$)}F"@![ETgJJNMH2>%;lφb?/-.i/">v83pGmE*M/Ɵk[UiP\)Mޅ խ ?Luٟ?<5,Ԧk<9/֔b! .Ae)g /oVB0.]U?(o.В5xr@i.Qԁj%RaIn v)Ji!1rw~ⅻϵc͇tqjA2%߭7ӛGI9XR3Խ ;S._[Z9`F˥M;+ IP>9x׋Wu=~zUp 1HYs Fy8{b(£6nZ^rsi#P4ԍ3|~ix46NzX!#42 wbǣb97Y8+g],i֦\ZfZHXxwʯ|Qxo%\rl9fUx&w6`4{|x_߽_ǿ~y3n޿~ o?LS(uQZkzʄⷻ/?_U(yڋX ՕA E\ffTa>Dƨoޘ-O 0'So/Z55ηZajF\|wWy>oA@yMEA h /@)+QG:U<:f+psZ@=h@qj7 S?'I "Z1Bm;v/H}+u DV4@4"s ޻P7Z"DiPK`)^;ud! I_h!x Hg&h,@pi1&穴 ,^)6! ՊS97Ϲu 8k^<ƀ}xiC,C5EMmy}DfQiHP*Xr?dNHb %^zK2^'ӧܼƃu7.p\Rx[#VPEp%msf,ecYD}74WN:Bϯb\_f^02syW۶ C@=#3*[~6@H._I^x&Pg*3L P3m_)-o#-HxD6S""LJVcF_6 -U՗]҆cWX>3zC:7[ v93$Z)ȯZ@Y$ $)$#V 2=W6(友w/|Y}'3-B)^(L= jѨ*o/> &lOg~{r:S*2sC7̦=0cp[4F392!|Vu}Ywej?t{ȍU!ȭ+@heD2o=;|߮_&79 |gKC߼VӺ/o|Wo<xxK0D F`B4mi0+CggGre&L7Z&ZqtAE!8m B.o|V'ETkUAfHY0crYDmp;TjԶ ?[#gGoĻ@n'/o79O]/̳*wۡǎ^Yl~ME[Ǜ68Vr 6 4p H%U#41 ɌY,cZ\L28ll19A!>@;j,z|\ڹWKK}^9W^lS;j fikQLf-'3?FpߘTHHj()JSVӪW!BvR]S!ڱ{r&fDȳ+ "+`!BIGBٕ v= L)(!rm`PY\謏@FrBf"-@rCs g͊c,q\ǫ zacYn«px?䱁/#X\N2j9MlTc by^#=vMn@MZ4iltD .fp .+{w@)]QH:U{91{Ȕ,k2& '(jYQc4ЄϽnZvFa^J|濵7\d;/U9ʴaI$<-1e]V /}M~-%T-I)J 9A(FƯA]v!+ɂIdL AAMsM[T! ;_f-|c0srEo|{<_g8g̞1Oǘkuϡ |Č3ő0ēbj8һ^m\;*FԋԶao'pQ,nw"[mېq/=tRKRnKWr@8!꾚8r;Z$̯U yCyt|mT?{F_&R>$Y)ѹZZh',6oC#USU]U];Ƽʹ!MaB%7엶?,}|Mgw ՘w:bdi{%x ޴ӭ4Y'$NPk>J$YL9;PLcގ'F__nv"zB1PpX\A$HaaP$;A=%119D.5Im:!6K(PmxYѭ,"t3:g'3b<ԾaڀnsSV2Ze!&H4ےs1ҋVNjhnjVYǼ[2`ZE0d1lcS, m $\%%0!!zQc(,<6ykY*xg<N¤JZa-v~^Dӭn -z |gCRtRE=Љ@A eT\)VMXåv,w;xT>Gptq͵ PB^ k9#R"%`wIEdUTafXT,1$8"` (O AO7cEP@a/YW!""c FD#鰐ޡB(*hPD0^A8:GBˡ\[*dvb oT1ѺhCOu `+W2#*8'HeFꖑ^ssWvt.ٙϙEr<. {B$9k.G8Lת,D<$)^ǩ%=a|f\x|0MKs=E?_ G;,9F";Cc)(xN7<텁. 5ÀP& -,InEW#, ] wGGpXV S S%\g$U}|濻39=VL%f[)ٌ0D`=lj +N` GWTE9L4}OnFX~|n]LN畳[ S?IT_휣HmbRu?1-C iFV0rpT(Xi' B`OeӸ*Aˡ6l6WFg?eu .*kuYZyE#Xkl+9hb*p.ui$aYMlxF+%H%H#Y>i83qMA{ y:KɇPBmTWg4/el5KQ4 lSvk[-L:[ݢx[q}N3 ƣAh2}*S} [Rh.qA &cp0% 2NcʠrǴy@(*tآVzHXQgkhSچB& #9oB!gh՞<#,RFl, Bm5-Sy1 [|O•]] |%m( Ix.m_aORAzEqa+շ12}4 ko|6}1yhURkIP|0![R[Lk;џX66ec>zR$xnɭiex[n$l *7xaM7`P\o."TlJX,^ +q5h"_ 7s^c`VN޽} 4hWApɂ⥉ZFDh}e]RAPvFKX2y<<Dq"uxcZhl߷1K x"ˡKgijE͉xz1ߩ׌aU៚64l@쮶z毙wfbBT~ST7M~ST79T7&Wr\52cfrT7M~SY~ST7M~ST7Me-*ro*ro*IM=fHCO;5guvù@p:j?ӚYjl%ʕF{SJ ŠPn?%0zGE]BFJ%V"W0wG"UAz+E jOft"d]Ktp~xxSPhV2 $\janUe:1"Hx,(3g-2٢q_EFך.SOVt;RUUaip1٠g%OEnH)EԒ)E 0sQsQirx'5`k U cjW逰 ՠ4 ij[,} *K>jwtz];k~;7ǬI;\nU-Y@X5dToK-ل^K6aK_Z\(ʞMn7.x8/`O:9-JMU@mi%%MXe:[p`1<'XdF8a10 w5\ʸ l!,.\^sL(QmGs4GA61r`]jZ D h:3tYp =̂k҉Ok6> 6=5moR[%kRԵsz߶%(DhTHCKK)FD_As^bR@2EL BІ,F|@Xk#WOD @RDT:Ix>'3dDpJoOz1wy-&#'XhϜ9:DA0]`h^XieZL3v¼S9zΦ:i _0۪^;EzA9s!Daf:Q3:vV2ŽR\ #-I0E^"~tҭZhˌVYǼ[2`ZEdt_C`buk"aMRb&]ԏ'7#|uw>T7.&' -j̩A^[vQ6t1r[Op0|je+$otQ<#0U2 V0bGe+d4nJr$͕۠qG`ܥ/ b<;Q}0/fJdP4G=XX.\><;a?:ORxۼ(, jf. إ˦MOW ̕~  }qC.xD@!nu=@ˮ@4Ulu9邯v9,M ;QC\n+skƵIZַg/-hRwl?+2]enԿ3;@KQ? RR,pv qI#iUUF`73vGIy`%ٖ2%kEŢ"/Y<粪N>7`cd7X}dMnjX\/`%pSzU)\DYl>#؀1./m朂*C@uAR<*r3 mۋfy+olr7p]6~8-\y?l|lsR#ȶ4Cޭ24ʝAVg:$-Jv`™RhGːx;|IQj=u~/ 5h0^lXvFI:rJ@]|^5غUs%>Z#+JL~Zȗ>-.gf{YR00_б8ԥrw2}V#Y#4O{?^ߏ=y/hǽIv uOVf|hއ]?o[ގؒ{ݡLyM~k 6(y6)P|%T>!*`"Om5w(k-"aiSPو)e/Y7N*ZϺ9h<+b'O'l^ڥv/\\SG C!Et:bVHA)Rd0F+Xһ $/s3Zp%9U %$!q e hFBNglfnϸ6[B|a|ZL{ r04F͓hht};޸&.Ă&jQ: Q@r(BqUe@GL:s*PE4Tim*$&Wq"5f\xtZٷ9=vrlutkm; &p.WAhKg EAEZϩ=96Eh[| `&LPhbQeG&ժr Z>bo?lfn8'w1v#6[zDzCu^;e|]Q8@ DYl3/'0.`V[)1dlUt ɡM1B*e|TsŘC,ZآDɣ 9IgWEhen(Iy.UgqKHV0b0:{N]D u"g1@3ALgSAjIxKY+VGGުR o0 ''xIӼϗnaJC32-chfy,hX]Z/hg6+ OIO]$a2IDR45R4RyF%#B[9YA)[x&̍(L!f"z2c YQB^8' D:9w8 Zȴl}kyNs+ay^?)br5z&-B 1zu``gAxD–N5g y %I ]ؑ(A•H[ :k "TNX&sԠ& tZ*@0kØTYBq]"Q#2cHkQ5뵷yNDo[2g7]ACsf_oxh|8ǰIX"SXT00v<^ރ :J$^ӗ[Vqjŭѫ"AG}v5]cq ?Nǵz!D>D-|܎ O^\S_:j6܌ʇ׌K9~q u?~^ Y?o, Cm~OƱj闫YFD5X{~&GD/Vd`bf@zQjozbb}+u{ջ. tdQyݓ;<OMɺUcYtȂ jqZ EK gf4/3W2YkΤ2+-tѨ050C{ZwE+-{㮪V]Uk+/ w~ܕG 6]Us틻juwU\z?J.{]rW\⮪Nv]U+a@W]ihg?g8_Z}x_wodGJL/ d.XmֻO\e1dFgUq'W˂WL=Z@jSߍ߯o&a?η~ M?cB~ǿzJ_m>uz5=h:Jpãk8қWRs9R\ZH bt]e_̨] ̹%NjߦI{Aݧse|};PĻeOl]v/+B~_YwT\hrbk%a`6?s.7C -?z ~-ڇ(}<Ҫs5-K}TW5EsNA R!: YxJP% _G ZGnBhE/d!fJLcћ7at$73g͟m6<1UurZ) Ӄ)xfQY|a;7i@n=4.>?̏F׷ӗk9nDŦ}#VzVeWvI[fma3sv9=O&WֳgǡҼ4<Oi3b캹*9/~Q ~qVkx lut\CwY:?#w 5x;! QqC8!+5 y#㿧 F+:0:xIE0)\23d ZcW a=1 LeB 6`.&me̼)2>Z0yQ[ylN_NO N}VtN#FWal;Z޿ۿmSϏww̒ߣ畖CktǼϚ+cm{.;RmQ/Ğz#cine}[jۜ8_A2Egs6p3qsF!UyMO[Tpw:$9,#\ZCX }w(w>7<L)#}" EP`W*b;DPtS. *YD-6E}Qɡ!Jem}ؘ73gO> > fr҈=,[gF:E[ҩjFM"F) S)G:3e~D'L3^GL(=, G[pD}D DȐHg,WAy}Vy^i%+|-/n 1"ES/be7iٔV]`)eO% =Liq4P>=Sorx1Zd5 Jzf>e4r^|^,6d3z-AUPkm+7EKwMrx 4v EE_jdkI],>,S}|g\ZPJE(x"5E xC*cf0&e6G%˩blV/G- b9G׆fïռЖa+,y% 4w?"3f_OXF2ф &poȩM M: Hp8G";^/w6_PsɆ~\; mgN(X(nCLSKo{vVmL༌DmK`\Y]Ymڄr+.vb'i D϶# E9o.\v33"MY=4;/ڰ@[ݢ63lY9*eB6H` drAJ!Lm/dTjFB6VfPCYjЏ(D7 aW[) IE]Qo k55ReGZtjj Wv2 s @O @8w&DdJ?r]*j8-yxD\ _^|.|qIԨ+i R8Qi&#QNv6)|l#W"gz&(aD\U"*U1cNKzR N )Ԡqޡ&=kQ{ Σ,ӏiZi8lS/%RXZES=1T ]R۩S)}%WA`R sk(j &QSƤ!T 0F+HzO*?5ߓnޓt6)8ף+5WlٖSmmw3zPrj~̭C[rݙJ'ʶ#'\9ġNrS|Aǫ8X&j 6ἶ2HI2LDlr9u T%X_6;4Udc2F.j2D&L M 5‚_ܑq+ZTg96z!}~yqk 7E$uTA@>$*2G) jluh1 o sKT~p5ε/VuI gJFjf)iu^!\XThk()#rwx˷˄:l{#P~3 s1?sWAZbњ^9[ S$ea_X|}0u̟Ծjv<}{+ds(?" ]S%(#;s7>OϏTg5ˑ7ZOHt6#P(%p ''pɧT~^/#b0~=u_u@2[%\wߖ}{<9U@ʌ%r>#U+r2oh-v½ xyrJ՜ޭ9Zͳu>~jr6òrv1H^sni?=-]r! gvXA4$TH\?y0yMYfO(!G5F U;h8/fynr6QlieD,F38<@箯ƈ_=FCe[tN%Ŀ=߿8Cvo|:᧟?毟ޟo|ɇ?8{u+0>5K@imi|z0|_zSi ϻaa pϧAqϺh\rM#3_?UqРCSRrֳf\(9q:nրPCFb7E/1=$$Nect[xU-S('N3O.I;"!h +~lt_/ş: ^&Ts&jVACbjwVm@ULVwFS{GHXբG=$-;LYS*+uD@\AEA|*%P{B9jUp}18>~eҋ:`o,V /d_ʃI ʩ٥_xe U+$8 }(pUp;zpŷ|h23]gQ3%ze\v}qx\qL9Br9CL-SmL%'\BbL &2R \erhWH$pWWT8 B%vP*S+[]e*\)LW\-ZFp|W sGQD~>~/s6fd>~lrE 3&\u{Md:B4ъG-*<$mpL)s-G%'?ىoQ%@LF81@O\ AI+,G BPSn-W-<@W0VqŢ6Kx y gI)*p #bȽ!EG!(åNqK n-ENB^]2{ÂN8zo6CߞGIߏu|yqR}Hyؐ\˦˯kf_|͂:L:kѓz]ІH \TY1A [KE 9Le2m6 HY}0l]˥v9=mz.?0|).ղ1v:-kY=o:m_V:pUY=kanIiM)[wC^T9OHL451)srD!j,jR!W-oH]v Tt._Iu3| Dr.ٌ"j8f&_g1.Ya\.vGPNnF(=V9Bl 1. T<Y 8!pq<,mCXk|=zܪ.a[oV>4z"~|G!FKu0K3JLi}j4%^ 7B7''$zϡw\ ZSӊ)|'Zǔ ŭ :9" J ->RPY 31 d>.tLyp >`nz ou9 kIX.l:w{GH)viuմq9 .ڬ njb [Svg=/\њO軙6:TL C!pNjb+όorLSh"J*種TRxċN쎇:꧴& Dubh7Ѐ`:x Z*($UFK8s=gc]6AP8gzH_!aT lExA )RPeS=4$M7ER=Hg8]ݬWU2#@'DLaYv\;[6'6mN_v `%\%4M*GG Z5%;0.Yt!U%'Tq:$+yH(IK!' ">C&\:!Ck+iew026\P QEtO]5iJ$toK,Ev:˧Tv90\j͎cȉni O[QFv cU -cf `C ëG«/:v&og]0|5 n1:2Z3v>2@ϥz^VS\ٮzޏS=O ?o[ځ7=I ׏+tReVNE)\q:cGgsRH91% ȣNsajMr&CDb&D}vȎLRk 酜ٶQT(ZޭzƞG[3"o9'cr*TӭŷΙ;3kGã$O]KW0ּK\cAṶ̈M}fH =cYޤ\7v7{> 1a}+~7%I)I ٿ|u_,Nn:9I9X[R\p ~沽HD?Ո"FWqO?-c{/vr99ߝd>!`vy<\; m# V',N&aja]W M6ɤ}y݉,dmK,ӳ?/Tus<)ix龹4bX~aIY%$ބߖ{;\4@|,oqohfNC -\p33ƶeٍ&oX]{Qc`𞰛ΞR)yD6{a9 $*PZ&ǗR5DlsQN2wudC =MCFV W_ܮT̮2c E|t92]-6V\ PPz+sݛޞÌo/ͅwo zA̍AzgO ? G-.X1?׶M ^{3{%Rm3J~c}Ȱ|Z^Hт)l1[.;X3=gꅾ}h)8=Ap'՗`XcAܡ0TsW!U(M3Vixu@/Rg-ךROmZ Θ F'-uu@ kaɡ%C챱֑67 OeASk j}<_EhQ&*h9~j} >-[Z{G}pjeb4z)éBlў+o|EQ4}DŘ4NScp@ :g(@,J@mrpO*5GcZF}B sg}+ݦR; O ('.ʦ Jh5I(*2,ɡ% E_ _hGD1.Н*N)!95^܄j4sW;νoO%{d|ߤc$Ok.}:#ONw>rkUIM]z-U&_?~lEGy)UҠxZpұH`,a ,DY.%X`$QBXQ(J!$HIX4:$NDV^ m :%*I% TiX횱V) qƾttႢ<]F-# _OuOo`0<ד~9nr>uJfD4# Ų\ ٳ(9YV\*/IvDD02W.FvFðb.M;7\huT,HGHx'Cw"K($iA=dDЋ<*5!DDsb]H"ra1rƩ׮ϊT4b1S#)tӈ8!phehTPQL &5#F2炱&ՈF3(oE.<(`)GG=z G kbl׈#~b=oZuJb\^d"b7qjLZ(qQ \ܠ"Op'$]PN/B/>,Rz㴻dc#w*}'[`{iEC\xGM2ޕm_lǂ@ qÜ~_@lmwi+IsdIV2(UǖLD  y+Z`.˅- ɧv;9()z8&E9nmsKcϴq/Y"iҳ @F%0fNjZi6U]Vsj=LJ3mKos]\~a;m0}LCgK5w_~9?_ncdM㼲ȴՖlO%Jw2y gÕ+;﹛fCYHaI6tMܒZmTZ EES.ѹPBڋYbGqOcuqqbbZ1+%9o/ Ӻ1u;($PX;3e|r%P3+)2VAo>8&.R=;_^;ytۉc}Nfw d0a06msīW TF fZqCx+@2c+_lx645NR&MU:i%ӛy|4n!={FB =eYJ='9f&c}4n4ހ Rt/-Vor-G] H ~\f}eUL=|]Zk}ꁒ"=vЛX9jRٌV$kZ;[$}X%,`3Gc+4,m t}yq`_`KSwMMR 4%Gz3]R;b Wo!9v=vف[>X5rk;J yH7/vSLe,_/|b"SRȄ)7)3EoKӬuI?{TYoK]sr&ykhГSg"k{2E&mYsAW ;ͮUӿ/Xf|qF0dƙ#;a!e|R;Aҽy^%c Ϗ>\\ziKTA$HMf7ytMGH Dw̐Zl|%[QѣԆ.kH)+[cAbZׯ[a vi+XinXW7|PI}^ c Gfcq{+!%9CJؿ~d٘ڽOO_ pf4ʵNO;[K9qu9!΍[KK&_ђ8jι lb8zoӢFK[]ܜI%jޅH;괛lcZ-u%eiζ89bU(ݻȧg~O>t6ADyוKxq}n;ThT6u^RW+0+!.,=mFZy߽dac0xI{(z#8B1|o1B8H?WeEpkl ;ILev׹0w'UfϮK[81_X1ፋܥLWws[dxl$ tgG 62' 0w K:ڤ[[X30ftsq DGcƥ&j8aE[3%r߯5KjufȶFO\H&ўAT/h'Z!źv<(5Qwr(8sAcPytp ӆou<',qYqAohMk$+k.BoR 0XA0*.Y&q"hQedZ{F؆HvPFEX;wVHcWA:9aq_$qIa*aC7ZB,*\D60Є6FfdӒ.qlQaH"zRJ>6g[ P KZ#q2M!6B ]R%v??*,|qu0 c$5K)(Fa, p+\6ud&eƒ @JM:s0RǡEZDR0`)DZTd"w"=w/:;~  v:5aŮ8#hd6ta (V*8>6Mr25~)rF}Y`8P-) z 6oi%'-Fx.`F9htȀg ǐvA+6j1.M5i̘uJHNo0FE*Z4e@; B "n,8)#IȚ&>Ш"8΢A`*YS̘(\2A~R*0qpUg dQ&m7j@36 ;dA [щbOEX;fH,Q;fVto B)J݆VMC ,"<*,e"En$' B`_i^Y*3ca(VHm:(ya#ރg~BSkkZ-^}8?k13G}"TB+[D3Ɨ"&ԁH!ضVSr@[ǨCZ֪EA"euhvUp31FHvQDWaA_)劶w ( 42t@s'V)w0[ `}fAÈ\Ba=!+CaKYMvK5%r$+ 0BܠE#AHOQd$a>EFT]RXD ( qhc+km#GEN;`fg`,g-$~Z,X+IlQd5X" ޣ+3V|NaT rC)=5Tc W2+^72$"%g (UUQh ̵ްՂ>?Rŭx?xR` drJ fځVi ¸AXJa0R$-o}/*J^lO`a; `XoQ2(cƞCo(kTg}! Uqu(",0eC1̬oR#4$ tA%`RZ@$إ֘{P]-vE@x',Bp[jpӥza"rqt:[p 5cӥ+dC.^)rR>Cpzk)]x/C!!(ezͤ=g/_䷹وy 8~I3oK𻏉Kgĕh@ڧs/3<'.gk!z9B vIdS𤐆sTw׬n0(PFExR OTZ&bx "zMYoTEmQ9I`A,8YҡVRUyIdys?DZPٱ"Rdi_o_sPI,`kolk;w!0?IkգwNzlt>W9v );E+"W<۠5[WAzl ,y"w!,]Կʈ*\yVb 6iH"[aV\BtYԿ\G5rX \")Xe)Ube1YD&5,U;V <.--aC{萸R;"mh-O\+N@"^uypA4x"rg*W*%탏Q{erc+wE_ .0,ϘL,u%3r)n 2[=jQg(>crv~ȏX8?.}kHY~S_k^:]\@RI'Iܩ ΣO-*5cB+j4Fk=7[=uɏ?\/^;gq+̅4x>­vTz2z9^eH$.زwuey 1z(AEɧD71'ٶswT^nu}=56HBkez*b دj7x섍su;԰-jЉMǣ8:鷟޾w_?Ο`f9p\feJHcF}2!ލ~C6i"Y~^Pֵx{2gd׊Nȗ3 <~ҌbEGO̯O˗R>B/?1jho1rY9Sw˸bGwEh9nH C<;n vgwmy=#&+C!v'[] ݩbZU#Yh;EcHX8o w1J*,E*S,JBE_z|3i:+HϪzAb+pm,D+FdEzk$%HA>YX¢] ɁEK坵&C\r㍬Rx%}(`W@|KR*)c5^%AD≆Æ^g^^ހ2/yhd^0/BG ļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļLļL/yYc11/A\w<˚v %7ļzp&OLisA3U1ª e*}6% \Q 6N}pE5dOP.;oݑvG`]\!fio^z}q&OP<lg-r6GS7mZMIJX%2bJBtd+Y# ޗޘ q^C^zT(}>X|͐=w9g_?0o K30֧'􈉟>HKl=T5\UzwvܽOg}4 Mo}@&7 Mo}@&7 Mo}@&7 Mo}@&7 Mo}@&7 Mo}@/ ߍ} UA &V;"I 0"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B@"$! DH B8H+^Pdw~1tZr}P1P=C7w??pi?xf~j5yKIi .).=kx_&EپQ5Z62DͷV? vcq,MY $6ˇ{Y R1EDɖsц@5&S-;uFĐT+p$ߩT:*\( Y̮.ϫGi/3w-%X 5B2Ad9LK^[qUY@wp>UBW|"<,1W xlDh@\YchPڧoѠBRD#(\<sq, ug,h=&sp;BqќB)EI+ɕ lј+וb@Zw 4Ws> +o?rg6|7j.M8S-MH/{ʰc[Un sZ⪣gQZ}@) %nVdg;V+7}ÇgLߺķ&BN{gF|IZ(*(DR4fq^Hs1&ي.u_UߦoQm^/VBGV䏜j@qy7λ- ) +w5oW};sʷu %ف}ƣ뺵jx5Yadhfu͗;ra?yvv\tzդ:=.4 R Ԧ:M1F"c2(\fiYG|/#pq|D`ߧIwjvU37Wﴟ MBvf5rwtK?5]7fz~}Ʊ񋟍,IX>ͭ^4h@/gv !:T&vIl L}olW&qRyi7\fCDI>ś9)b<H{E!6eX½`g* KRnY{GO%,$+8CI,rd0K,{>hexeYe] a]!4 sa_;$ wN7"uO}Ӏ'/vv@j>-^FΒv|+NOJU&3AwsH 5l9~էA۵|,K5gr)JǤJ6dgʘKE?~QiW{mעg 83tP{9ECGvp!0⸀F_v4~ȳo݌E#ў;[-Ôw YFS ,RLUP; y[.!9p`\ͬ͝iRf6օkX&|-Ӏ2Xn\O$"QawhV7y=B.&7Rt XXO042A]{wOg&x1m.8 p8.֝a5w"e7snx~?cL"hs4UEJQ " +uJúהۻpo `xyeن9~´=a\0jj]&[".}r][oG+y݇~ 7{@lX`KcTH* ~Dc6MWTW_uW]F]7$^ #q St2X!cb"+}3#gf0}O($B`$_}q_뇧寭 OV]yEs~uw/c+*ykgl :{sXsj-x.'zJ\![Ua7OΒY7džw1`OfQƥ&[Gg:^S~:P#?Dt,ݑxbTGA8RZfeyv d\:6t`dJE!ZXٰ`F8J4B'CW _}{*l,~k7r\lIJ(}qf.(_/Va'q;dX#Q!H8Oっ`O)z eJL:rAb'2H]eRpi+6|mrkU9z[a070農8ސ⍳`i[;G{zwчf`@g_vNo͘uzՙtay!Ȫ|*Ͱ|V /I=(jT /`GT+/JO g5ԭn"D4RR?s5czGNŭL%OPJs{HG%lUI! =F8GRfHG <8{D~3H ko{.waZS8vYEL%VR;A҆iأgBG?mǦ&z9%_ڼEZ{^)FJ}CIYYMI#}^eJ4%I5h!eMbQ3H4`R.T\Ƞ}JgoLZh^sRk.hއ$}PXDz(`إw%1Z+Lbs[S a1.m"㌏@'D^{LLM x)ocXD<6(Rp&,}sz7P$KL㢒) %20b/0P21c%Q>خJI-} @q|wSI?]zHD(i=gB*){@0ypK" 2he?yLz{%[O޴;<]0Wm v 3+]^QSX~`DzY*GZQK= 0}!džkS"~p5_E I^A 7;$WR6}#,E"2v%\F)O*\htl=YVan5q ] }S$H -4gzW:_.~2Ukt/\z^%TLzbX,)^hHOSb's 5EⁿgP{u9 k@X%&eA+UP;˹& SvKcPa &>P<28 lBqHp(^#1&Ef?OoGZHA&Hb4׆s)1p242}]I`Hu'!9ÍHi! q~~{qhF (>^px|>ߍ횣Tmm}{F[XWKv-]v5#que&bѸQ4eJ> o=kbsw*Vj^:->~XΑҰ&xʣdpP1>kMZMS{˦S{4Y^/|ve.wj7]ք8.e7KRD2[9s%b jƈŊhZrq Fށv׵1gTQ~vA%3ˉysR'_:0=q<+I?_73DH $]sVRI;Zw7V$%pԓ,-aNO!9dJuBP0g6h"^]!ρeWJgm]\y݄:dt'F'S iXA9M9ZK6vZ잨~Vm! N!.JuJIꝧ)/QUG-έު,Q;S 9 )E.ECΘj3FBJ:"2J.JNیZ FR HFe&nXT4P[{[l1uAe <4E=~Cb:{W٘F9:f Ł-6Y>dXRYGXV:@3_((fSFe+|H܎9ByP`dɺ'kb(QSմTv`i^j tT[ n48N{Q撵9&E_!E"LȐ+bI%[K#BJ.P;3[!Tm[R,T XjqETE&'(@2j#KPTǰbiǩ+U}*m!3%1Xӏy4}Ӿ7zX2P^ޜGsp2 M&5*hLIpcO=1}ښO[Ñ1-Z^8&rBF J0AXIdSv 1 \HaAF)Kkb?O~94 <8=}f<܎B @TN)XrEQUB()ViT!K% rJ@p$B@+ji3!5WSpI+ !>vfqqwFh3ZMYˠhb~͗˙hi<%kψ%!SZf-X 4IiשF2  Upu .x#DV 3&VN5qvq+%q"vU[]{ٷmaN1=|&fiwmV$-~Hi.P4[(ؒ#x?.,Q<<CK $1vr\%~mp,BfXS/,E]MN|rф*m]֢` 7̛izDŽ2)D4P!TXL[ve« F!6rSqbwX ()k^e)Nef)z۴OSl'S\nSD= ..gp?.<ެ8}im잛1v4 k7esw2#1_}j8ȈDF,J~(=2b.2+hk{H\I`p0pUĕp(pU}("R;zuՃL0\=L\/ WeaR.fi\jW9;?EϹ4?|[8ፓ6hVJ_.32 ےҗ3PR f\r&@D_``@šldAHWBUWC+V1v*R ++: ""z\imIJk!+ <*" \i_aRN|p%dn?=.jiXޥ˴o{(XlfFN *NH^hÍ$,A22c@mߩA#ɍሚ}N+#yꎣ, <^ϟ?/o+rqvx^ 7ۺOS"ZLS^5=]9z1Dw **8V[[P?fP71ekg%t%#`,{+GF:fXogtV-mS'&#ҭoM$<]#`dxbe 'Āe+}0Ne8V (&BT)aK\3@A?w)q!ᡦC8dR1 IV(9Z٘8;Xykq,CE[kz:Q!7Qӯ a;%dak^R6yxEK,X& Ve Q)KKM#@L%;mR$ep~[Q$\/p~=.m'T B$j#4l,`\ӊHBR SVXObi!ҫ[Ku}Y๬g2҉aJPPhPB>("L1)li$8ԵB+#V R|~vzC?'?g~J{D V89jZ|"mt)Q"$9Oּ KƗ.-g0&|{2{RfZ߈B~S&ڸ~7 2&8%;: C8rqG+tn54rhT-K6iqsq- H?ӓ2jbw>G硋)\onjwj HCL,9Z(1NYV&'=uF5g%JM:R{3~5ySl0$ sz6[wcpPDSmu37ykI-\źff [dH 1(2e'og=kѿ9׶*Vg\꺾"9ou\l%M GT}=&ʽz̏xo8R^ıyoTќ`g4O>ߏN_:y _ߟ|:J6uwPZk0gz\}7?ǡ]%=iM|Squ؋__L.z^Vv`ԣ_˷4jڛ5 wh&{=]څ&ݿo0y"bCG֝ (wS)򢎱QGZf+5(J.+W|Y_f`t6Ӑ,,79#'$F)NQzTׄ^Hg&}B;9䴓<ƣeQj.v˶d;klNo2ټoliyġڈC:]"U]ء a= QrӸ 8xs}3j2#F/mu?&0DqJsKop^. *]c- -҃uN}LuZ'3@/ۥYm7iJA#dI T,OmĘhM%f-cY ǵd&K,^3Wg>]}.{7giлre=!z˿M\c~ݫ>75Ⱦda{ءN$o\jIZ׭`Vr/UDk;O溜C 91jy3cq>datvIɝw S )唥BνK G X_rE+ ېd6@%/br"EH0hkMBmL 55FܨdmˍdcdeVGd$\nkˡ5&/^ìiN^ʁ_86_ٖ1ϦV<v[$ǹVjQFy@B(mJC(%leΘh9)IG !f1[j٥hrFST\@+]Fَqư/  ]]f|ZbMviW@χ7N̈MCq ʀQ9F0.,u g6e@F9ZQ` ¨:nnjmBrHmFَ~8 *f[PԶjw v1X =:e!-L.K1 9sJhwZY,vnEҐr$-:@҄5S2AqLM9,кMڹ1qaQۂ͏=q1po_ "z33@%B5ZeV #28ϬGlj6˴!ePi$8X5\pG ёdB MZ9btَ7g^'\Wզfɾ(Ebf jG.h2%r3"$I]`I ;p)p-xh  ln*]{N[ Zw<fwEm\gBڂiOlx)oK1h'f^ c &ԝ4sQ'2sKFgɞIbFNhgt,m,J~V6e<*6sl=c{,?/u,unz ^usNȍsQTUZ?AC.oeEJWT)]MJos0R̖x}y4 dE%g{WixʧZ;ԼPr| =z]N!KsZKh > @l>5gݟ5PjL\9Ջñ.TZstAk,fN q?uGIV4 Bm3vgUeIo1b_qr'>XN:9{Ce61x4 Q1FEGI8o|19kw6YZfa] ab 1<ӊbV兔%͙h"Ct&H D$TNEji (ys:JJoBTw&ZQr(MJ,+KbFeH >dHtKe)ڌ\e2^_fEӾ7M&ZB6݇ g㪄]z**,aAv3gKrgKTl2Yh3ֶR QP2!((i> mVP ۯW|:YNv`< J%2FJaƜΠ(s"eZQ " 2K\f7e&`pV%˹qIVg>qM 792(Xʍ}Y:.8y]]_&3:.d^i/Y*&,k(&vG{*0 hc}% 2*ϝxIf)2jZu*Ө$ZBօf- Yq)֬qL,BdA2kMJv+®l6{QȰܔh8H,N5/@Fg}D2zRՔC2Im)q).).ae..gp?.<ެ8R0I˴֖0禱tm4v݇HLW_j#x2楗r_|yWD pMAAJq&4}JVvyWeټc/6w4L%% O*%dR#/D{%4 ) mLk̠BS.Oa(A(sF܅mMg;S#L [fbk<޵6ۿ"ӽkH],nσV#y-ٓ [Kdre3@fgԭV: zJgid:weCyQR^\F2"-y tQ:;2Vz*5QCTP+kg^$ gȫ0ǜQ)"=1T)tUUzF ֗Hk1@Ws>9Bb[뒒$~Ow 82X㳛.C!C"RԵ efȎR ^xzGЅiQ)]܁h)Z+t. -bjGǤWʵ{+y쑾,cm1Fs25mhElȀXN]1?mi"y,ɶe}$^WVmnMV< 4L{0hDɸ  YddYٰ7%]K>hD#L=>Gަ;ؾYU؅.)\ ]Ƭ9M4*id8L٩:`'xT1@$w] 󶩿rG?Z5kuWu(eBRHT^ + ) obtWc|DD[>FG*/Ûo2Xs}rt{^g{|[kgy ^gf iL~Z$ʯ| 9s߭,90lg?L }WʤDHVyeմɄD2y,]Mtņ)ݷCG?wo+_^ujﺽL F@?_{ %J58Gmx{37d#\M9]/7zg9M*H/WQBjҠKx4.rh7Y$рdqFA۫r_5y"&oʼn`&c) Oq,0%F]|-hȮt!$ZhsqYd@߂wq2&&Br['f/=Ϳ_LW?|] ?i}%G?zߦ}&Y`uFoB>g/;,7__}cLYnF}3Ǎ*]tH3eu&q NW$^V U{k/ĕ`w޳'v5RJ:8)0 3:!EC+nOFs(b`^f居UEE#U1WSbcb2QWRM'W`w:R(̩UVcb% W@XN]9*~*pUuኬLWzpwEvцvL{x3*Ò.hˏqt.lj? R61Sb3TNHކ!Džُsght IB?Cx&^\Uehˊ&-8-xzkd^}:96/aFyzz3g`<>~1ǯ~ЋaQ^E~M _winu`{%s\))kq2kĩb=j%9㕔KRL)yiۉ}xun/_/]hs4LЁf5%A7cLLH˴eŒe'qe0ȡlAf̘Y Ҟ۔XVD2)fguI'r!JvK,!1&1ۧ&)aOd%5ewCCqm ͛# 1{pg<+yp(k'_[=J=+<4Re d#`#)$$,ofY9DY*!w#!t&!zc;[ٲ m̨ek@LI#y([ʼnz%F!.rkU8H,NĚREt>F@JHcڙ9V:r=# yPtu12/>1 7?.FPil4=Mc ~ȌTz j58uup;ఌ d'4mhElȀXN簻cxGo ,eʪ-Uudg) 0<$Q2V Ydd`ub- N,y7^B%㟶8,[ ZySbqcyR.Is 1YsiT Bq&.:`'8JWzʢeL,^e*v)~n˫):rG?Z5'XǺ:L)kj9P˹r{-^˹rRR \rKr{-^˹r{eGSv$AZwB`C2sN%PVcT4fz4\݋ՅdBpʒ7h1eu`\UpY-$ KD<H.pZr(AEgQ%$]Mִ7yBo&=fD#Cck$`-|@m$XځbrEgSW~|w@RV M@Bx)ќ CKNJ2a1 ,%NɡJz 11'H(J6Rr) !F0)c)QV81R(J P`fC$3s'A$ ӛXܭ9z]9 т69i^gL0M|`zĩ?i:Y՟t6c#Veor?İ1#:O 0bZhs+"#7Xa'PtW`p`p@Fg'Wma5&$ZQY Vs])BGL;변.$^ǭ~Ztӏǭ6u4()At3Kd\F2"-y tQ:;2Vz*5QCTPN93/MBAGJd1gTHO9Uy ]pUŃ8 Aq^KP t5G3,$漵.)IT cμ~a:g]ޕ$BYN߇!d'Lf"ѧ,yE9gU%۔(XbXuU5pH*2v5MLd/i׃n yjOwgP %kFJ[:)q9RkD)Һho=w6z:ibѪu0?Th!6FŠְ }އQWx @[x!3 %s(iק)/SNS軦 H}~K ~m{tE'm"i#KIߠH;Y !uхDrwS )6Dv[-ux8pW7hղHt{eJ"y҅ݢB^v QŸLJmdqX/.2ԮFnL!6TC-u ^ftny5ysü%M~qULùoi,T/Gi Ikl>~u G/HZN4<ܝ6 'qb, Q%GVY`xk JMTQH~\-s KkNgs^pb|NIJ|ӥRr>I 33abZ!EF"ER,"a1Ҁ ^' a Rşlw7""zB1UGaXϒ$]X *jdGO&FL.[Օ7ب+b~/Gpj; n8J)vZ>҈uӈVO~O3:9 㷖 7[6#]iwJO_/3&Ze&jn2"ʀQBkL<:cXF5RWHXGJ& 6!`B!XQc(,<6ykYzotwשydrZAU}S${|tsX];#)*AO RåG!4,5ب5\jG8 rg,_$t>HA&@ǣtSd?\+So{51"b"1L":aX3SD/bMQs;:,aZ}0Ь)(XF.*ɮFX 2p'ugEa4-8cL6 _ʓUvaJ =n~Tvt|z~м,O%f(+C |٭M¿/a8-$LNMRs-S=:cFHBFK/ g<͈J"6U378u??y՛/N7'/^?D˓o^4y5@3J?f\~}1KU|zp g ng j4(2Wh`~. (zS˾bƨϻe^-& h6;%(!63zN򼏱GO?w1F;J|`)RA7d+ 2`[bMCt{G[84蒗m_F-6MV[Pv7$tjp^}Oź9Z{[LDlƙJYhvUmU~o=GGL=?0#%i_&}ĮS ۅ(kcL0&we-*[;ox_`lJa^iGjd~rVk/1[,} *71Z;Et'[gum Foo{K,-*V'I'N's 10~ n<)Uj)5 }KCR7xAp>\݋a\{PÚ./[UF} 9p?.YEə@8+ /uV Qe_<4KO uXdgVh=t&k[+p+! E 5]QV!,}\nlYEPX:̓jf ߙ^NFEG0A׷YU"^co&x=?@ddy g!_J9̰,Vɧȧ0h.sLo~>;Ȍn[_d6oK}.`(6"?ˤњb"f!H9IཆRD!e!x0k5f,`ZFL&ZgUHGzccXRަZf[)Ƌ8y"V3GǑ7(p&:-41DyVS jw 3DQCR]QQ8faHB9Uaa|4B퍝 5U xašUW՛ztrUhWe}JѠ9Q+ɕKJe%Q햺J^QyuԷX %4r%L K"!L< dD%""P2z-XJ&|e>jbdF/p 72jdV {ӌMPppQe[5V/Uâ]Հ /A1/Z<[5w 0y"HE/9XqA&k6)X,U42Cr |j-$& fh/#vHHLBDN7vV#-ݛvl j=j-إ^\i0:TKSXAHBGG^Rk s`1, oWBgO 8Xh5>eW1 XGm􌇽Kq[7#i2 ji-S:u4S(0iA~{jΠ |S ($^2Xos$,iP΁'MFbRSψ;YGm6u%"La l FZG4pm`.8$F傦#=.>.n ۂV|[ܱjXYNoknRQ1ڏxAܿSV)f66qNL+|Dk8ڢ0Ǎ$@a#hE,##64y̠3I`[UL>̱0-gܞ;t-i&۵fWwٯ\Z1vBY+0Z* b{M4:F/'+tk!*[Wv2ޢ|=x5:U幑a2/\]^7|Y_A؛p^۱?p^Ru7Cmo8Du_-ٝe/=o8]JN k3.hî$R'n)D%cD'H\%Jr+pr*qR:D3bsp9\.' Wq+ɺZK% ppLIv\ \%vJ*prWOǂİ;WhW*qKɶUrήpMr!1J>z>z\j'W MUbXĮq[W˪\q\nNĂ:)0"%\,>HgCSk\L Z &ş襣4ɜs ə$*< y&0LCns=!wf}o6E%n4\aJT}?x4Id+SG޾:>r\td\ b\ ǹ&L{HyGR J!J,d3F\v"`2gxW#x|1Bm3g`8&gW95?;xe>${/ͽy)"IcYWa ":4#QhGB`RF >_Ưݝ7x;I`_Ø" ̿.# Q0YV|\\or>,&ԁnC \*A,b<~1fYv&o Ƿɖ@l[P?tY/?~8d]7 eS;!9\ae#D9mC0 8ے!/}q6өLύRDUP;t =EY탕"F &F,B0#+F"`"RӴ7b1h#2&"s^otL[7C3%!m wIv Ir&2 -L9z9wj%Y(f' /IׁZjc.(0}JLG8F A{D'a:9z1Ab@qbiLqo>nq篚PwkY,Y}.KJK츤>iGl9: 4K|;.B%*F j < 9 Ƙct|3;kHA\Kr1{\[޵5J4a…'СrBҜ rPKd ftZ|&P]C~9W9v .6J!۠LnE:˃X1XjA{O^d\-K؅DR)6 2qșD̥ukX癧;V-6Zj˻UO8!XƆ2ta:ؐ ^ΰNi{tLHKfg!KlX 0EN)gX#0=Ċԧ[%@"-xdHQKp+LꦌyL Pƽ^`zq",U@mQH9DŽRn&x9I6Y;5rX+"D4[v5ݐh] $:_#m}|yӢ5zqsτ6<H<]Cr"g\\!r -%& }ƅL &%$SĄ "mxRg K݈F8@RNA,Rg2 "k/%%\2N,s$u6 t8i"Z|jV.>V--A #`ȃL [ oB8=B{1!2`{ꅕ[ŴNFo_H,~2^zoWOլؕ|7*s9G7lqg҆FGd><dc5|Ikнk+ފSCLSQL*G%eFEpBJ$b-1<<#9g!YҼog+@A --A+0 j"xa^#E#$^ٸW bxŠ0"qiFpj LY9a Ii5\ᨴSS#i k iysJ?Mɻ@Ffj;vunQ[4xK=2J,(fqĞbjd:`I,D JFF]Lj+V Y䴒@@;ʄ +P/ϣ#Hx|h7p$(˝|$x>p6IV5x)Sʤ"lۡ03,* ~%HD0 AP$@N)'1}d(`Pm)8am1FD$Fa!C*{`PTР`Ppfuz2MiئN8 -9D5@H҈#뵷VS>$yL= @` EN.0=|ʕkAPy06l]ߋ,".5g0#Fgv6:Kͥ_\i6-4IY>yN1I{߫!2AZ|zMӣ}֜ +pRp 0E]JA)L&&SPga_v0$l& F.*lFJ+`An0w(2 3ᶌԝ wٍ)IJo-7{|z}.O%fPg]0;/6V\ {(UrgΪS,uTWޕM. Yd͟"WR}K"`ٴ Ci$.Gb|HMÐajfYރ PJ+Xe?*w܏9L=ƍ}C6 d/F02%ҳ 诘9>٨ڛXc o!*'~ r7W/{\_]| L~{7`u+0 .5 oz.yJKOcJ|2J?d?&.m_7ʰ|Pg 5LrTϿO]1)j #0_P=b?1jho>4Uuv ~>nB@qݵ/U q(o;sNCyb1_$v"S2 <ɘW42ǀ)@8D;JOqkϧu8+ƍVJ*2JF}0pf, { yK:.nTUg(Q|`»˂ImU4|oeVXA\7c0I7^gu5oJU_̕*w>EͯfFLgbR)0@`7" ]/Keh6^L55}=տ>g0p=+T`a5 nuD9hEy:|iw1{_7Hoxz ~i1F: lp6HF7g)UsQ b_ڍ_v_sip8>9'q.;m>csu=ʛaY B|O@TA<sM2w42e%Bej];}2GaFPs龻lFDSL)V.b)[{1T# B:#QHYu2LwZ D佖豉hj Φ2-j>Zf'M%V3wF\"@0 B" 1,,VJ! jg-Mq kwܝJWg|ҫm]ݐ\\]~}+EON0$w .)&snqDO\+6Ki4r%L K"!L0B9Aa:bKQY !ҩfJF #T +Q&x#32x'DPR]kΦ[3vU:ӅqơPEŌҳMg 7k.!OL&ޕ6#bbGef{n7*|%)@DžZbZŧ'߷F#Eb 3V7iPh `זUg=LvggJT0,nDx6+(*gg~ ^F]jT%>&e4(ALt`ֹ̰gٛ8;M dpDDVOV=ئ ܾanظEd&sX╅6ϨƜ!%̑i*G-J3K\~C r ꄜkD]lBu}>@6or!T$9Chzg ^˳<.'D _pbzbqZmWT,K٢JrE9Fpb{0!c|, 4;RRdZՌ&iLƳ4!)3Jy2fk!d "JCfcAf& i"ʖhj' c!ljEb8Q>2ּ*:T&R$ȜN@[\>l;Ͼ.<):ryۢ,SQav/Mc)T41y8YW_jň^N}"0ʋEr&.>TVsUևz ] \q%*҂9w*R*SWLH`.]qUp<\RA7WǏL#[YO%E?޾nTQHr\9jlJ,ԏ (:J>etP `thCd6a'')U$2 ?"ϼ ~8Kw_&nz]9@=}l9_q6j#͗|Hj^tп0ó|z~ˇ#]nq)e5Fkn7Z>|5h7Z|.݃bj+F+F+FkњoFkњoFeyei,Ӳ&d7Z|5h7Z|5UD..'©ͥi>w"%֐w郋(Vs7d^xd1=cڕ0Xx<͑ ?\s/_VIt)Dg5s<~KKc댜lf`d΂696|<8A{J°z,}cδc8me)ln>V LbuX6=j'/nfnF/w\B٘h wx4 liyJHϭI8+".j\+0:yBE4A+%> X^H˅ql1Xͥƞ=ظ,u 1)cr V h$Z9As㹂w!jq}7ۧj6j2ۻ%M{ 52r\>J2Zp10j'4,$ZLJ|" fJ:OԢc?  mj/RԷe7e' 6z]hUoa Ue|y_=,~rm>^)'p vuFqpG~{qlVWr2)ܐ:y`ź\-rŮ;zEz{FFT׿=nkiv &8{>r#+wMo]HOw tWWUeg[f=tuw]>0{qb1./<*_fT40O}3s宎atێ7e].6󜏫^5[&{ZK7 s?f<4~bܹ t:Ē]>NObqѹ;0y4Ŝ~җ#fѦ:̦'i10Sy7'5=W]t7HԚ+NF-}>ns']*ƬEE A&&m]]ŷtFmnOjb.i=t7hiʷ7wͲb!٨;7ܴ#\9gl7lI)6:%g SrACfzxԆ3>=/!occ>{0<Q!>r+ŢK|6BKkڱ,d Jgq1G)(,=V]pj'R|?&vxуKq4&`m,B f* d$k@81"r;pta&9@NVLsIu,1}Iu`:x*>KịMIUM&%7:gNJ+M,\NW0l{t9=_]sco~+P%wVޔ;X&xZﲡ.kN>i%if3t2e؛Nd t} 4gke/%bz%KHe+|)qG4ă0]x6Df:: QJ.&>Va4,h@ؠQkS Z]h}k!mlH(흳hKioCm5xSg^ꃙ%rD gAG" FluD幉I8o}'3ׇ{4HŽ5mMzpTO_q>] M]ͨ]YcWWM<9E/?FGQ)s|g#/taf~TĻa\^VEu05".KiTD!d~BtHg}p)i1 SO5^KFrQ1\ Mbx O]KȐ!< K:*J^@Q8g8ȂKLG̘yud!@z#rW_0ڜ⌮7d[&Z..5w_ro\j+IzϴuqؑvVgڸ!K40(Q4A@ke *w^4u`RKD_ĔI/[@WS$&jA =qe#ԯT ) ],+ c.'CjUA@t ! F&sߎ=E7Pg6>x~{'`A__+y%꣕B<yR>67 JL澱L53aSB+ B/pO2>Щ 11YYL8@QפID&NJ[1 myp4dЙgZOyx)dЫ(){.eO3/Zmrѩ [5XwmH_,ټ"lhtf&0xpt^dɶ˔- ֣"_ji|[goHeoZKjfq[_3El9c:`aC-(U=Hi:-{x)=A|QNj\cDcDͭSFD0Zh"o[kcS,#8Ը>*QR ?6!`BC ǨU1kQRp} \KLR5,3&Ζks=bΒ$]W #\\q}ҧWVTmYZΫ* *}+A,ׄ? ab>h7p$( @h|Aˠk#R"%1o.( J3KD4$3 {| " B)dFcH;9 `D4 RC+ES3kӞxTY|=pE۟ji.)Cp(Xm:s@ENh5!DFYCҪIDzD%Hk` ENtc0`=|ʕk [Xz3s?;.ק Au;]ͥ]#r&o)hb$?XN-^=q͟[.z)[GعL1,v8is 0E ^HA)Ԍǡo{ק ij_fhVt*XF.*+F!0 ]w>c{( &5G>;3;uguԆd6J0zu6_-]㭒bq)۪4S |yF֊KXBJ՚Rf;Mu4OL'.{\^[ovaTmtR45ChUK\l[jnF,*> l(?ہߴw':V ZjU+9筎hH/ÑKUO`b'툟S6vl^k*k*׽u_pwK~xsw1Qoo`$_Ɣ֞F͍^3J?v~~}3cNx["z[N27gtjw|vcVUr`ģ Gp~ڦb[4Q^|vi.h/Mhޮڄ(LYMҲ2E6^/Q_;)@2DckQcVr@l o+S8֥phT>qD(aDɢHÙk7 LI\*9=9(Fl:qh;==ڭNYіqU43:oD`Mfeoϳv@;h.j{3ϓKDb#M{"q3ݤNbf$r}ҧ3;lzS\2*y|\!;|QQ j7Gd4z)9DB`ϵs0t$Ah1B$eJF ,G @VNq飆M*Ffd =D0R-c6q6[|X%-̦B]lQ¥{3iB+R+4<_,iL7Oq D掁bRK(.dyg 9E, RE#aHΞO%$Itڒh/#vHHLBDNsb/biǡXm]vAkW\iհRSAHBGG^Rk 9f0&˯Y1!fE;$))'8D{`iLf{MvdG0E̦;ZDZ]R,2 P@4Z‚ky `,҂X!E A!Ɓ'z($^2:p$,iP΁'MFb+IfM-2]x⸽#~H{ٴdWH2E\b4^9@53L`<0"d<#m 2kF]܇]<{M;c&,c QS%.ۏtQfGDYh(k@8Iv5IJ e QZK\ ѣ1WI\L\%i\%)Xኹz9J#"a8yӤx}>EO2~@=2X?_kdk J2{'!&p:Dkvf;3HquƖSx >qk8wdv}T>B76ykeW^_MѰ_[n,p?>Q\.tqJ;}ɨGthJ{ s/a%̽0^Fo^Qx*sй$&e5}=JhuoRLA( L#k#6~@_u=_iI-AV8Y1y Fi(XkZmjOn[Ԓ}.MgwjB?ݫC7 nu^6^»d<ݻ /U{r6+}s;UxwKMl۟DDݮ= /<0Sx` L)<0Qx` L)Sx` LڅS,b)<0Xx` L)<0 aۏuF  \8A*`G\,=WFK 83Gtq5ǝd͑KfW"H M33(L U%;qk9}9Vе3a2*}0kE'bI2 :IOA bxVdtUSz}0Je;]?uPծ;5\fJׇp:y4-t~!KsV ᠍v٭^'bjoZ(S-yJ}xnS.{|sI}s?˙^avIr!z0@ W1ie_c}@*$T_.~9Qk"8bBQ >j%" rm ^ےj{#sL(QmGs+:2,"1rbrX+/B"I쟃,!yrG/pfTkJĝO /gpYGK^Z7$QwG1l]3sd8:̉ML?b֓*$JY*S /yS#*5m ';E (9#e,+$X`T9F$㑅H>HԔ1тFQ4`pH ?W]_#"r]_fBvj yZ)v2eSfa)t#NJsG͉fȐ#(Ʃp"Le/S.oHӘG0 F1ZleF͝QFGʁ i}t(wM6q6{q70 Q\A<;U'NS4Y>g9='ÙJ* ɴ31[qqŐBFT E"jRr?.bxA)8@Vh欵`E"M%.CF)@fjaԼ' kUy@a|YB7zf\5aQ%f:ˇĖcÅ,_[|ql2@5WxV0Ja,|1f怱z=]07!-)x5#^h;ʡ2F&0 k3>P\5Y2>LJAeh MْatWdr/<*YDw}XܿxG[uxԙOtجgx p|)vwMFthY][oG+ľlrR}3H,dX*qM bo I4HYY4ky߯~jíѷsKSWNk . >P1'T2hf( ř@L ːU s,PY\ޠX9gYu%6; w–=g~w1R! +Ԯ:OU mۃyPc]ExG7GU>}aMb l&Vl"S>ܓDH?=ii6ۢj^ǣĹ")b :r BR>qU)Y.utɜε^֮ZA.ĩ׏9V A4DHFv6)|l#W"gz&(2N[qޜJxb|(|)LpA?PWDk#]~.٢,qjT}Ͽ/ۑdt3H-X_h +NBakͬK/w= aPk>Zy)ʓ7eϴEYPZ5a@ІHo De;c#h hNh2eI%Ҵ\K)isBsޜ1/1"yCZen}o4PO./*mpKәPM4p-Q`"U h?p絎iɹ Ҫ$o`3MB>UOy4#1H=/"0(gCc3KҠTRB !Ny2\~r%#Z>SG^ҕ,sbx+x|v(#1NkGㇸN[{z' º -dvWp`hQEm[v}Y.݉l /ijO}>(ƹ> W㨉h1;f^&MF@tLs8i_ PhmX@|)Mjܽ R|SfdCnt諛}1LGɸ7)O,b~~9U 瑎MQ_ucȎ~o_J'VdǾ0ٜOuido?5=Ct7P45Ek{fxKm3?Lz{{$_~YhM[mgyΰMMx=Xq&z\uMa)?ӮeD<7zn^RA'vIh-ΝT2FS%^B"0oq< %2\iő>>!_^PKf-sڥo8@(ϊITZinUT66%ʺv^@.*oUw K>.CuVkLb( GJB7g'Η,)I@!.s"7,q|Sp+N~whW)+N3VC.20[vMF8L*R4F'ğRp遺#NUqۍ !y.tH# 2Nj"0 7y)e9;^p'5`3%}_֊3_m8|=y7{,JCgeYy˝T|"Iه0LDJiPWghn 0 u޴?i#=1>- <φ??`h8lha)i˸Wnq3CB4b@KAy[n ~X<6HpCI"G|eB94Z h2UiD4&JOqiOu8+ht޹ā3`*H!D0&Y1^=^5"zYɟX4[iT~LeUeht鋚ͪ7x2=fhRuAۻfEג)E #@ :#F.`k U cJ#68R$#Y@T(*5X@U 2߮Zps|ꭴM߽sedZX̵Q*g er+ bamv nd]%uSBj`R)G&(Cu,7șaصjd39kH$|О#:IhMvoK׫ D;t'tDSɸׯJvƬ0o^3֏wzAhgbDp{׻'(rӇwF9en&ȉR4gKr)!tca;y,p+g20FQSõ{-&8* @hŗzV=X|_>9힐Ms{s\㖡K9Yz52mu9ON9  OAs߁ֵ_[ 0%l!hÃ: zHkmc )uHHJn,"i@&G8=ηRT7xILYE,gXÁa"S.(S/42d-¼5z{&iJpC| Z|~f.Nٓ~3W8YLomt;<7Irγ;}[xO9?0OjvP@;뭈:m04(ńAzk`qTYfTY.D(N3SjZYVgy|Xr{ڙq_./;0W>AZZT`HEϑ  XM<0}'v"Zq':0"qFpj LY9a I@i5\ᨴSS[r i!I}@`K~)ɻY5*D&-VwS(y=gGiwBĤbG)Ih?`5(A(k`3/}>EO'y]qǩbHh,0G6 \ep:ӊIpݸ~Zwu;8ٳx\ףּ%]6!P&jFlUJNAA6 pְָ1}.B]5@hD≁Fw7"J?s1H b \ 9hG; CZG w^/,.eà#?_խ2ZzIN~ԽkwԧCfM7ȈڰGG% _k~HI߾ӭꚮ9ϽQ T;;iŵqx1Zoi]T&AB7z7>@S71)bd9j:W nW]2+M][.ƍ2w%0uYշ:ҷR׺('Vק=4\4rs- Y4+ +v-v݄qaFEVw:[Ͼ'+80id_F׉`Wm78\wCӪM>$kUyժh(fulPb=t\z_ji\߮$77}obV/Y*{:6~={L^ELTQvT뾹Y1T,xi2'"ڰltM-7/#I$43abZ!EF"角G`QD1Ҁ CzqXCug1ұY G/@7u-ՐpT<;Qr~cSуKf+%;'?US4J%eѲK:%93{Vؾ,l: eZ7]Dm ٴ0pvfdC-n4{W]*}VnYi}})rL_nVIdx2=f>]( 2h{y {rjnьT(A_ū5y:L)ϖ1B͹,gv+kϓ!&H;{"h U*>Ox>^9(3&Ze&jn2"ʀQHH u ǰV{e4P G‚FG6!`BC ǨU1kQRk^jF| 9k;=I\[+|5NoOqVkˇHUC]O^b] l.ʫ: N Rr&<8(VMXåv,w{x >B".#PG-%#)SP¥",c03,*͸Jb,#Fpʽx= " Bv15!0:?1q#tXȐRZ!4("bD Ykģne7׭ښK Vb:x"`'ba"J#ZM?#=6qy1<% :X)8`b<«Sai3U^#'H-wy!J?lhgfy1PDPFNz}_'אhs!YD.8~-mU!QI9&!5Lz4`&Sȣ}(\_$C; dKG#L9FRP;6El:}kʊN[(kn!Rr5U\se<0W4_l 'MvmJARRO/iaq[ 9L'_m~^q [HBr0)[xuV5O"giPySx5_^σ%jF̕7e5T휡ɸ<[|}iCi$.Gb|HWMÐajf~|Fa0,e`3vUo4nգNiԦge#|6j0ÔG9|~\zRY:;8Vgrl֩+:&YuT3׻?)_oO1QW f`\:oz.y+JKMfB|2J?eNx5? Yon)bj&@]0g~xh+2Ww81b|OBl _e[Cxӡbk QO]O:6aܟl\wKClZʑYMsaM9 ^$v"S%zd+ 2ǀ)@8D)JmqJdohu8ƍVJ*2JF}0pf, { y:KZNrv*5k6>y6wzc:\;=s1)d[WXWa5iáZ(ڜ-rm|SS0&gQfza5->S 1F:.#A2 ef!=!-\pKkɣ&}}Rʍvy//WE>fۗPV *Р]9 &j;2}vL)CI!-Ug ]ks[9r+;ƣ|I2Ij*S*[ƣ-%RC )a]@3x=hhIɒ|6L"8 AKHB)*5霣IW i5wTtΕRS )/|2#Hs(R98D42"]pv\|d H]hcSR2"(&' )gR! @taI^ТmU@!C@YUb`I6,@hrt1DmZն28p՗cYao6NNx}Ylj61X%&VgY_ 3??Ñu1k2h6*R ,(U )Yԅ"$PMzG9Sdd_>Ue |g,TkdlvdlUaa/'"϶TGåyZ\;u+Gl_Q# IYJJx0+b2f"A6Q@Жj C:FUMm]-dBvN@ B`)ƈ,|Gb j7ǂڡvg頍L:(+:eSXOڛ`鬳NGDsHɓ"͔#cMgJ`tq52S2q,؋dxDWcAfcODAtD숸iG3R%rB7z`s֛?xkbAN@QřcLGN+iN{scDlvD|VǴl싋1.ʎ7x R>RlR]ňat`&t4Vb!pX8<_C_r|Gdu'\1sFLz! g?P"g?^y4cU~]'t20Ms\9cJs+i) 2t%2OgUm L`U*9Ns<|Vquf얦oݛno|> +tmaD?ݩ{W|swN[v{sMEhݽM]˷?콖V8fn>s̻CMu[X.Wt*>y 8߆\ ŷ<>Kt_d_*_FV*U6dnp:[5ZVN)^v2ۍ8!w2p^;ѪX-•࿃Lүջ?/׃slcz=#_Bfmj>\] BvAl.ns?86@3xeB@ 4gwϰOȟ@rBfQ AiIP7y$'9&C$!ħu:ꎤQ-uڧf\=}NIkM1ը٭DҪ = &Ӛ!TjIz-DdҭLV6{funm^)^?6{fnm^6{Z=6{fu,6{funQunm:"vnm^nm^6{fJmzc6{funm^6{fuJ93k.~FvBn<Ĭv`^` .ĂzL]w~bNE œʹYs;*|N^{QY't(MuOBZ .hP V!^鳰IxDvoΎ}1 W߾ f lOSludf]aMeJJR0*@D l@=TVڒ(Ҥ)HC l RPbрS"#<)e c$Z76 g ~gy\;(Nv~uI9>)VRr\WteR?%wqD>kN#rORY2GfٗWIAU]+^q0^ChP+bl_:̊ݣ2GΔK)8/)τVg9E  :heJ{r/'X1G&׌]YUǶYެCE,F3$+TK2*m2䘄hYዒYޭٮ׸|L[{[\^;s? Ǜ׹rMv}n CW-Twج] k꾥Jdw)Љa9r" GGdYsDa+ߺ5b۫[~*p$eze93g_/ϞoNb逴M9+Tz>U>V'V֠9@ɉ1ĤOflXŧy]W0)|tE[z$xr=ܘOzi [$}Ͳ閮]_H@4׮y2_m0N/+p~>S-jd2wK6 Fh2\7Xl{b'KByȂ30I3FKdVҸ)xKXJ1UVh566F* &dAy5XZSjd{Hr Mmp[FBSJ_Vʵ| M@Cn'O# (ۍSNxY*` 5 ]L1@<g`H+4%(楝2ǫT&y91?ݺ 22~0zdp\<*(v7_7"M<>,Zp)1@!C`Ru..u^+sϾص jt*RT0yjŠ=@@J "a H@Eq^8+9͑iitc̙KS&zܤul=ߞKKjƥ{];;z!v\pjQ| 2bbZq:zq գECD@ֵEݽ)QMQz^!hj09l%YFY _3&yM6+885H|ȼtǚs\j,}B,`>/C}GR?z؀2YZ5kzZ`P*Nˮ aQ(-bK he!STPK:1FB@Y$Z$M1q6=$Kɽ \}jީ$Dڐ!a@xJx%qdmth"]W~zȉ E7z9;'/찹ч:wJ}p%!-B]2zMQEsbOB6B46VN%qY^q`^mB\ǼtӶY\Moy/a\x Gdtp1 8FVYYQ:Ji  Pzw`%%RPw1`SYP$-R94^[mh\ObDŽ+3Og;w~=;%#789gPG&}j12gw-C%Cۂ!RZٝ F| &a^"-^93k)x.g1&CS>'UZTQpMƙd:=t:1?i\MYĻ3U: /\&<?6|U2+eV.X]wr7e*aJ/Kk}EP~-zN=ˇT=ho[|ɞov;?Wʃ{ȝvkLJΡB@cXrG`]-v똚=xIKhonMxRȱv#2֤i1}Yf=^]-ss 3dkC&uC5zRl~I#{#=d#|#×Pqbn}ڽ77>e7]44ٯ?qItDK#׉α?0U̿2?Rtuir}LGp-=nh-EP[&-m޿pOz3fk4x\o/ɐg:h}-U|Kc|Mm[0eUZ_ 7.vЯAX+7ܼ^ϼj&!у:[ :e} %,B%u:)lΨb_i>?֫jm` KO L!fTDISV{Ǣ[ZtNktaUH$U=lzu&0oTRwqo[}w.~d) yc1r{&]O/d젽={/11έu3) kauI!7iqp3*^>ܱbP|xx#RsXϸJ7FS6:$/&A&H1%BQS\%?{Ƒ@/Ig_ k{O$:v*1HYvE##RPF@3*r/K ;oNK7f neJn@-b35QcM8As^VV";n@\M zk_ke ܪCݽu cRINtp`xβ|+W2G:?icp09sH[G‰ leP}ުQ[!u\ IUp;oB11  1b1)j$/׻cZ{{IO@%u a2 O[5S;|SN[}gCugtLeCs)巖<َ 1A"C#u6]w0ggE1*[2`))ypXF5{HX*&L}(aJ;f2Rʃp-1Kȳa`a.p(VHxh}Ӌ}5r.l&7JTz隦T\*QF}&Po,XN@AFWTB].,#)S@RY|Vf܁Y!ʓH)k%=VK= Eoc6&C:31I#tXȐrPZ!4("bT YkģorfئN8 -9D5@H҈#뵷VS>$Z3G 61dS^qc.0!#, Hu P>h`9Aju p1Qv ~SsqEygj8|k.yG4LuY5'p@dǹ,ݘN/?J>I}I{ ΅yUQ>ew:l9F"2@S)( d`2%x>#ϟ-`_3 8+:mqJK K +h^$ Ttqz(23߲ S*,̾^5oD^O/剱LӔ{6z \/uB l.WW.OnT9L~niv6{=]ٓogV bj$|q4<=-7s[q\MKg3S,'w =uCQ*|A: fpq1Mi\+Ad]uce #|$BF\>_W̒Gp6 zl٨4*'>Ǚ^;~N޿}7'O޼p:y'~|w0R4n\F-^92J?e\}}SwÉKþ-n2R2Lag~|py#vo>- P@/Z65*A59o/mr5n -uT! 0v@3i[M:kqe;J@>k'"@06P:yE#X^f6zC= 풗G-?θJI屳6RF #HFΌE\Ӡ`ݰ7 H\*p^iՁxַtA/UaRi:T=Tg'rJwE_`w + X uŬ+*QKIU^]=u7\z<@TbG<4zQW yu8*jPWWW[/_Mrٳ)]B]D,⒎y$}\ڛ~hN nrT40ͿzM.Wײ7{cX/>B)02#tI;M> C ;#LebG&L_ӻG &uD9hkc2UaLƲH\42˧rzmcIM8ʢ_}:\.p j 9Zr;Af* 3ĉE' )&:c$jiEh^Exp!u9`t]PW@-ê*QYzureݺuF]%regUV+R^]Du8RW`!;RuE]%juP)+'zurD.EvF]%rJ֣+RWܗz9Jp.ꐺ'QW\ɺZZ7_ tOg6H ~VͻwG,\ /Fy^$TT2rgGv Jj?,6ֱvkdQ`qԅF,Đ[\Ni$e0۫M%LҾkq~(ݴ;>#_퍊tH'\ Ny)?X_dom_lO+C3[Ҳ])nfQ&)XP`.ZH`mk$Ղm}9&r6ȹQjS3cZ)"VD hlyc䬿:[xRy,.)DM2Q.ݭ~7[c8\\<|zV>/?t?c "K7`ʺK}MJT>V"I\7 qဨ_)ųb@wOǗ&'$%YPW dg~W`Wn@ 2&C%6*5gԽҨ}ϲYl4<'ej|>k1bYOʑT"g\ rS ]O/${S]OxL"PAE};dNw74sFj-F֨0b)?cD2AiS&R/5eDDL ` Xy$RDv=5FΆ Po Ƿw<leNa[;Ԟ?K{S鍽8ba9D3cdjsR*({ATD0wQ vTQh4Bk^2( F刊`) NGL9%x%zr|9Oc]ˇ=56EM~ʇ*+DHJSeYAo3 7r0!v9fs0R(؈$®&EoBDŽd&n ՌɄ܊=^@VhdZa0 b E)8jafO/$0=V)(rXro0-j0#2m6Fz:- m9\Na G?vӾ\_ a,tz,x@xP!txu+QRybdlH!s9!8A83yJ?QJ#Yr Zdy,oMjQ( 29C6T+%JpOudRK.^,Q{xHHjۛ}!I=aXӄ"K"fXPJe^ 01Z`|%T8 y iЇ Q"! #K9 ށel6􍑳^k^krꞁ0'`8<=Kk FA 7_u1h1Qo,ۮ]6ok;[-X)ϾT}Co`r&,*Jd1WٿWRvT_v$/ jQ;.u R-[xxs+wK5gn;Ҩ{L,c05)%ahnDh:vF&x&8;oLm`pX fgo&ALOvv8]S՞AiM5M7\n7$P 3C"mKmI>PDڒzӳ)uX[R W0wG"UAz+E Ce{ :5.vʖxCW;Ƨ7FC{<ϩ)x)KOGn{o/ʔ KStZ,fX)k& >p%p##D zlD'azuZ:·Yt<&Bpf?Qv~|uZ>𿶵/ -7ԡ:gءT&5Nh rS?? =:j0P6ၵ6rt!5*F: lp6HF7g5{W>Q^d 2-T Zr_?PJ d;jef 6K+VO [ XEd%gղ@y6iv ={#A^ec87㣕8L]+ ?"R~%)J)1mQXG7cq㴠e{|LZߣ:dm%5Mr*:u1d@a(gJo˜BcX{?$]tz-O~dgG+7/DEo+;ה* G gY!dѻ*Qإ 3ܝ2\ҙZF۞9QifzVW 5\AΕe.p`WNKRix`,( V0 b%wKvnᾳH؅DR)ƀm 2qșQG ]x`r6?ĺ = NQ y(.݅7|m=1Z㭘ͱ4(7ƥ4*A2ԓ+`{VNǔFH/I'˜ mL/aN(Esfͭ$GY{j{Ʃ,p 濥F8a10J A ׆2hD 6W֨A!=8|y&k"c |u*#A՛`GG5ٻ6$U$`g"ЏjgTHɎrHCRP Y4kz~]US7g47EI$G{QtBQ4V^ uOs\NaD( : s<, II'auy JsNm_\bɁIa᯦bo*W'|؊̌}C>^e@s#h*Ep (j! t ȡG嵓^)᭵/^Mn8~`٪Yb+޾yR~3/"u༈1ԕB3&(W\c\T[JG0} >zl;DX4j!:#NмĀjq2JxLBXxL~*z뜹g*71hoƭt+_~:?D@AG wAV#*bRD1E!J=5,(t#^!ۅ6{` 4aTrR[`Gʁ 4Z44r.INic!!=  ˏ 5pEi&Z8b=^- ]c+#jsߥ~ugiKa7Q)q99*EFgnGU8 }Uj_ѻIԣJu/Ƽ-[Eas&6-3;RT)]ٸߞ?߿El^ 9GO-Od&ʩ t9VOmȮ*[.`s۞Cqhv k-d۬ 7ېABjFHirr2.•\\Ae\}Z̈́5X}Ph )kmi9Ћ|[vdM _!bҀ٠)/oڜ2]@ލtjKS$2q?1?Μ CG+|;CsX'|t~hy,˅خ**q(Z0meEQCGxtIm yAP%:~PUuK6mE42%9Ze0]7T8r`TDCAezF <+[G`?XCmH$Д%RG D4YኃaJz{ldIzXGv1xM}.TРJS$n%hƘ^)H8AEV*Iѕ@>w].͕P⭀=*84s0J('j9fpGGK#٢k\swV Ӝ6Ai:m`%*6d7^ I"$"W݊0_&.q1|.=uzpJ W WUb˩///PycIM~QjNпH֦gx}U9B,S=zP nFt\]y[^x}uyVfQ9[ rm̜߇4Uk.8p@>>}s mvʙ\=y4}i7{IQUY͜庛uVFd۬m*PF|Y~}LC&}5CJ?Gt0M,؜k* A^v/??7o?߿>=yɛ9y5:,Uݗ9wWo___7/J|q>7|j: BMuaԃ.0pgy=]M5Ss#.^m]r˼ݞ7k@yy,!Ġ"QP|>[׻ /crCIuP U06PYgBDY挎}%jlԼ)JqhХ(C:x䕐MgiEvR8O`Pozdڄ6|/v*⪳=S2jNC^x rk ӭ1L+>iba\ﭛl2&yT0 1mq?dG ""[xmd>5 C޲*;^ucYwz[ww.5 }kuJ"B8K{E锨G#,hcYx}ߛ,0z1\9dpjK0M)`^?aA4;?MP&> v)A2mԁqiՍu(%gp䬬N! ^rVKWW*/^VՋ ~P%A>?YxmCuչ^-|YڜxӢĆnۯʃmPY_yW̓j*sy5*} *atUev ގM:>n\OsV/OX Qsj Fe)j}fť|(f (F F.Y]q3eco~?BўZ0K#:`D ԄD9Z*FOL๚s94|D$0!j %ՉKYXU2,j-dR-3E~%W IS$:2Kj]$L[nq[Z#pm?@@ŲBJ@dʽ QH4#2zC:,P~/hE4ׅu"L.%}ܫm/evKpqUw{4pJrcRY ll햺w?͝oGVjyF Q% !(&$gD a"2:Ṟ̌0g<7Wu-ifݵ>f7OIw0$|Wq"G3;B6ރ=wm:E5z=y\lZȷNvCffuFd !7͞*0dIY鳁̮6Xn,KǍ#b`m9ǭ`d &9le=Q3a"m@ ݧ=[tS+5 l6T=9̏[XN&,_ Xgs*..F_x{'lFfuʁb0ž{;\8RYfd?~t/Syٴ]7>#M^??-'#!iVT ZojEMC8gף9NGOrb7.ΰ8yn&>m@T[S3?/~X?fvXKOR_6\ϋOwկ'W뷦+\,z@`Nj'Ggo_|{R[8Hs:CFGw3XqlԊ}l|/W{ڳNn~J|L~pXVp`ZEk~& (%gvZ,3 }~틮r&P h9?z( JYI+oI5RjJ1hτ%g:EB> * ] (ag:A2Np gYD,hvv(=)ҕuJq5 pz(tUڣO+Jtut%95U3ӊRCWKf[z;nVհfG6ߖuIKkjp<:ͭgRO"p?Q:Wc"?⒋!̈́U FU E`GZV+EZtZ85 `}Rf(tߊOW@N4wl0t2UEK骢LW'HWFHfUKYQ_QQq++T4 `͇ p ]e骢4U\HWN2-ـIά ]8rZPPJ{χ%'bEcǻ//F_7ǡ'.'O}dzX]q݁Kn[6+, CkT]a;yA㜢gٲ`T.>\5j2ϗo~fPuآp-ZXspQcGwHa]˿:+Vr[֥?^ S叫֒Ws>˿v]M=|^;}QK/Ku{|[y.jzCu Girt[cĦF7d,&7}GW]Y).|vӇiXl`r6V2R;A1$^(bLj>?&݅9rѸ?{Ҋ̵{;rؑbq͘=$<3A7:E|on(޺]CWέ詡>10466|MG Ac~F~jqrUvz GZZY\_Od-XObȊϝ}1%}e} f24wE1%rN׶>}\|9ܴ157)ӭĿU.51w$TGSYIvTw54<':]KS =g9wg7[ƨ.?F0ҭQJpu_1C37ݵE.$N]rס W;5{Ku+ZC&Ν͐7rsz᣷=L^ը2e>x0!_r+x>l44PJ%>3d;\zt|ͽHέ;]ܡ6ic5T0>싌l[ wz- 󕽩AcwsQ$ >b :/!u1р?[#+\Ç"+Zg][YϘ,Im+Zʒ aQ_~nfʇ*~/t~in_6uQyZ좿5WUj~xoA(7Uv <|%@7~uЈxv-c7_ 7Ʋh&W4M[^"kXٿ_5>^M߷<]txIkln`u{s[^<xÒG֌;bϼy9aokTuqޡ[Z8 ;z9p|r`.0# ʑuc/oCZ;){u蒍_xd^Jl\[I#-Y'=\_g$ajn}^L_NrzWOMW)K^yd񢓣lK0WR$c)I\9'bevzM8| ndхG)O 2>$bH\8Òqq\ h0RL*Rǔ 9p%c`,Zy (6{BV,g’ `9adH8QH.Kw*hnE9Ir]͖JQ* 6E%&(*2W2@IlȩgƊsIV-S`K $"fh̢v(%(:)l]{M2n,ØɑؾdzUrji2O&bA0Cmנ U4u*I$ /xCa.MhkChx>bMOFZERZƝqE aCTN9hGҾq5G.hi t+mC uX>1&Ykם7[!UC!->ϱP=OmVh +/֨= lu>0: z;}mv*zAߣK)6Ʃ>2"X !$ZTOd`!.&:ezH%j ^b1JK,oyD`̌wBZ͋/$9z/aEL AzYx!ё}HI%^&ˀOhLr U)K& 0<8$2c4ւZh ~U*AIӫX(QrRPyUN+0 nc0mVa+TŅ%xCV$+c П!h) ^FFV`$ b}} )Sf5aJ%V_ThU&Ds"orBŋ EgM` VŢd2\,gʼnQ2u4!#V6JO +d3r=U& E*,@I&bZ"D^eH$ ε%1|OFb= o}AxHV˽V 2hhG"3*Yȩ Տ/XF~ S,0l2Hr$ $Xz;X 7V?]_gՂ13 cu NBz+}z322DQŴ:vMVgq1֋jwq/O:t*~v4ZS[_;<BΘUY""D1W :;U*7Nwyoy:b .L<-PCll~Vje"\v @elV;ZE' G_U~{ @ mͩ?l5V4Zd*ءMyOu7̾5Q8BYq~6^#L:1ܦ|b~}O7\!ʍ:`u2F)ގﻯ(;6KjhJi%L[K&N>r5ҕ=ɔ {JdVm+DyЕEͼNpBT j.NWR]#]9M0[ "']!ZNW:#+&!B& >iɣ zNW֮ 2z3dTۮ_-Ǔf5ݞ-´*f 2Lc(|ͦq{M ѿS%7g/>NK_p_Ż94:ym __Bth}TG+DBЕ"!7 ]!\ ]!ZNW#+uwut`w_tpO t(':BRywg,+˓+D+m Qjc+øNj 钡+Y*thv(i(2XJ:BƦBW>wutS%DWXԎpK}0(ug%:2]!`+듡+@;]!JI;_]{v2rC{W=4]Vp} @W]oH&}b9./SXP2?\0Ur~>nOI PBtT]Bޙ; xʋr|w2oBHd>_YˢkE$roKD :`P'?4^uT/tk,J6GpzOʟg[x,R >T5xHއ*R3zPDD4h }D(Vh$]!`ϒ+!f;<-}+Dٷ]= ]I!B?|y ]!\S+Dx QjEtut~7!Nd J ]!Z#NWR0xt%&%Rd dvw(yWHWF*EBtVd_tpIJ+lJkWXtNƻBNDWGHWNzf\˓YjGkWRwut|wm+kퟭ{y<]J)iˡ+wϮwC~f ?RnpvCk9P]]=tB7N>~_ɕuN&4`/Y2'U,jQ?q++9OHwCy-ЕF3]!`璡+kJ=]!J-O"T+]*th "Jˉ2{wpM=]!JEkWHW%tBt5Ut(!:B'+lOc,#Z{ Q*c+g) "c7ޤBWqwBڏ5ܦ `!ɻjT *wBzY]UX?lSŤOK$;Kv{朷Q$S/Z#rϿSϺN>l:aƱα+Iv ?PD Jxϒ]ϝ`%DWt S+@+;]Je ҕ:%d r ]!C)IvCҕRk]`#}2tpZZwB}{-Еֆ_%ZfS4 i@!l> q 1֋Л.e/oWV{p<;YZ>/}ɯufpkeǿ{G`x'BlfA?&/m\e׀]^6fſc̮Udp2z0/b*uU?UInkN Huas5ֲ'Smʻϭ 5WduA3qlDG*dԴ1tuFׯggϾ 5UUGx컖::mS\>_鎶ەiԎ) ia,S jXpZw_ Q: -W3]!`NhpYG^G]!]Y) O dս+Diɻ:Fr)+,H=?vBeq#]y0O+O.OƻBIR1/luC88*o7f0;CCm 7M|pH/@2@Hg_$٭Q-㝯w_$ QZD}yXgMYqI^-iIlI`/Oմ_7_HZV7p3VAQ7Yr3?5ŮϺm>V% a WdVjr<ߋ E^ dg>2x-h! , sJUT^VQ4`^|=_bU1h^s\so=0g&;i^%fά݉l2+A\3{5*tG0#nhW)Lm߼OݏPcw]ѭ]9w}lkS+/>S[mu{'6W'MJ˰<rr'KO'WR2zy8ptz^6|*?L'm,Mb|(bK!Zt`{Zҩd %f/0a:?٬X69>eLs{]nD:< 4(}9c)"FK@. vǐU:|g ,h9YtKq1Ü1`Kf~=¸P jv/?evs/԰)Dwy#i}8sr^q;|8w/Ok$uET?{pQF_yKY+B_ur>+u؎]\ nu^<>%"D7y0l{ q>- ^FԂUܩB䮬Q@Yت*+5p Xosg둞1ht-s>ƪP\Bxs"5*ok.x  \sw`8WOwR nWǻuk^+Nҟl|"5{z|_?Cnܐ?{V\`Տ` @$__!ey_j>$J%))Q`[m{YujeVVtltсj%HdJDg\ 5[7KZA8Tm#`̦Եbd)eRSڰiI9ƐƁڬ;5:݌iiu%+kgg*&L_O_Me% B] 6)g_QTh}7VÝ}f1˿krum:y<]{-ןo~OR:,.[ U Jd0FTP|PI"PMzWR*!|*Z*fV )'_ J PajͺqfX3BMa]|.-/Yҥlye&_FwL&B,h"0FAԒP%9,.QJI#%"B2A@Жj kb8uJ ںh!XHNarX`%"Yw#vӆ<n8`wyӨG&LdB:&!/XtI$9 ibF+dfXEGcMʂ SV>Z&QTL6;xج;p!'y6\\ "ހ"x|`Ѯ0 mQ  iJAQ$@BHH""zkl@K,IRq&=ӑlJZ$Z1"6~Drn\4KE個.x Rp&g%$t6C6LF֏*PcHx \<?qFK ؓV^ѓLX$=7d?RQ4ُS7HsT`VuZBx')񄦘a3渂3Js<ꍝ(񨾴Os4/AfF<H%&gUV Y i(|՟._.lVf^‡n;kIoj)U"tnX*j^<m-[6ϯs G \Ooy˻jZqd}$29 4M4!Y$1_}%-5& :&Wƹs<8kbT2W cW†Nfr(4~#hBD4D֜9#挈9"[a7(432(S,I>I ;JW0dC*Mtb{r` 䢥)SɡRU*3'aBl4 l3"xO<~<gvh;o%˥ZBtB&@A H.֛<'guFLT2ݔ#mcJrz)xهp ie ) ]r $E8ug߱nA0eԷa=Bu|۳2{o:th {$CͲAv-*rwE<AFS>}_KŸ#qڢܧo^:wGnqn]RO"aoN>Am },IW2e u)UP&JVxk4M%d@GFoG7gAS#NEh J yt4/s8'Wu˥jQ>/1 rlL)c"d۹~;,B(_4LS16H(k%[-:[Ja*Zi&HrŘC, mbLT" { o֝{紩VΎn7jsH&}nC1PWNEMT#9LwutE&5k:ݮ:c/[Փ߿vz8SGo.m2ingKJu?MTG$cTQJ&F'D**Z"4mBQJHM$OXBȚ7tN)-w2Eͺ?wǵ@Ql| {w-VlbP/,y,^)UE*)J˚>^pD.zGX`ԟ-_-/k?@NxX\^, |Koow׿b;٩._ԙ nFU._]yeG] _!=Jͅڧ::;c>ޔ\;2tȣ~ L.X0ti]E*ƙG ׬kN, <%uAE44ɼVDc`A2i樓B(E0ɢCv>!ht)+b V9ʒXs~ߗo/=~+XAp>D}\z|iw;{h˱,-A&R(?e4i Aü QV"+-lsF2ٲ1(G: !俅"@I(ZPI Ra>ߠ]-u;p~T|k/Md7睺``,Yz?gQQ<{Z;8A=Ɛ-XS@͉"q?'q3aο?ccYrfu~]Z>G'? v̮fhnl?!/cѲYxN_#y4oz LB I,D"Gޱ薖DQ:5(UTTaЫno>O~S|T8x{s w}szsP7gUo#TL?4~҆^{R&WlqZacGs%%{ѿqtۊB#wm#}.Zٗ=z%-GTB "Ϛ&B*.KZL98[ j!,Cd_Գwr盵Ο|htBC Vl0 H*/S}JPMb38J r˖ޓ~*n?d>=W]֠'+Fq̯xwwO|ysg[wBBw3^([Cʪ-kiJ_?\my?Fr'M$vEZe4::`Y 3!ov1 4 nγί#VKQc`pC!wJY4%ij& zỼ,O1]g^:P^:PІWwvTß.DZ4 X|>Iq38,\9Cm{dy`׹~my77AV=XmAPKlQd\:Fo( A*_̩LPdΦ,vK(^2hpd"Iv5 ?-7^S $p4(T-BĭDmr>77HdMj̙nV*IUq*AR淺JzutqzvQ Kj*,yC˦[!o-vij8p *}qcNlf_+sݿO|syqRf ,sMxt|2_[nvP6򢲙9qaZ֓ԭ= ӶnX{7vL,#Lpq:_5 =st:krݽ6lmyD.z+hxMA*y\ Y԰}&r]ޭvY2n5X.1﫛:>dv|^QSYL)fN'Op5m2݊LlЂ SHQΔ>GM<ps#8V|rkxZ2ʐZ[3R> :O^_dGTGRv#zm3Y#lfY'9NY.R|ˏoVOxx߃=xue? e{I~|p$0;|KƳ<T'6XKwwRBJʂ=mP2v `7,G~}PT_WlvP_l\Q{A yb7c7P-a\3[nSN|g(".ACtF(@)9™|2JxL /u/$餾/|߾88G~Ox_n`} W*J/TĤ(bB{jX$&Q$鄓q;k2 @0X*J.R V"LP{@HP^9ȑ29( b>`˫`S>PsGt4)wD Tsyf6*gD IG$K<2 )$7"{Y35p8'e\)9)Ę`^!{K׋A?\oνGpn9s07Mva8r(Hq&E-y_I̞W2;RVUOn|+qRܑ}uFbZoiQ0aj}9ox{g_V5/ъ>Fu#ۥ#;usoکi֫y 1z>5N5:v90(&Vd]քM]vUmjѭ5֤ʬ%Ŵz.>Ws㶆5gmPx޴SZ,nMu۳aQl8gnY7v;]iRNy9$9Lyw, !ɎPuwVʛ5n_i;iCgO5ew<0~4i,Pø1|'vc޺1UMduQ&L;0<>UdSHrQFGw<󗆫,{9Gν iBOhP"NXpz4*M1?X7z1RGЗӕT_lLh!9\=4'd -U$c; 9ԥDp3׏qWYA{ >+淌=½U{t[֚jEQ nGJlQ*SNPՊ"WjiOe rХEW eBltfI(HCϤ,3`D*XrĄĹ&k b!q:r"V)9̴ 4$.Wwf^4F^QTVبׂ"-iCAZO5ƻK⡏T%0^)/pFƕ"ZuɈ5{PHHHZ[qH 2FT{"zs;Ǎ4  :-E@L1Vz#0jDlY#br{Y_{㽫аt.Wi( A*i%;T 4iEXkDLB:cJ!,=Wz%$4lR`);M Zp9(Rr24ʉ4X po((}ύM9Zi@E9^z\(a+څ0=8ʃNBED,{'!4tHHO A=5 |xc|x[ *ф U6<APB& p^FNzA|`^Fo/Ė[Ϧl-xѩ_0_boxAyFE e2:nkkO[ٔoJpir ?u1P ;LZMQ[%Q s{Nȵ,ۋ@HSKB5ᖮah̲EE ?mF ~v>6ɺ*#[զX:{q#u! F>?f+u= lUtC%oӇu^8ߞy7߷ߞ~훷Wuoq $|i\ΔLFw~m?}Iu99Cͫa֣~Wp85[~Ɖ{O/O˧'MhZjۛ75l4%%G=m%?Т]jC U rcŐ[ӟwj+r$#2!ȁr.$ @3TZf 6;Wkc8̭KUcu8h Zs$SVA !ZeXGQ{*(=="}ᢶ_c2mgγOvoЋhaz&p;88?QL?nlW@pHci) <1T <0nk| F+:$F1V()5khu:%*  )IL(q_@L!kwWGm}{W+׏%MlږӷX[DP3Tz#7=|62Ǹˮi%8U!Ep죅/M~CO՗gBtrD$nRKTgXt&&H#)bA}"1MP )K$EDImrp43N*dbr{jȓmjpLk\Q>O8ݜ}eZn<k Wo UU3Q2ѳUxGoҍUurmjh3\JOקӥ;wWӻ V߮yl~R|.?v~仛,ϳka4yJ3B!m]l/P>-\B4ßߺ{}Y{z4+άCf̭O>AY \?J+N2Ts org-m\/ֳW[ulip=FϬpb"'BZ$Y0]mSGظD6R#WQ2ZAԢɨǤi~8N})H-h:S5SGBpD{$~`s E)i0ǔHWQs; O ('.ʦ)WAnP,^S (n*QQ$gAO!ZѰzHJ'S mHJ[YQŨ.-4{C19**19h9(?nʛ8ZJJ!i < -s-Kb{Ie\ҎF%RBrE@)CӃkB2زAV 86*muge,Ub-4-|Vp%nfrqC-]۱U|x7TuǓ/b#rv@Z] `ke0@[.#7D` ӂI=눑̹`E-HQ|]I*ʃqyp$Jx#fU|>_"%.#jvq<?bR]d"mbkIjp7"F陴bQ@A"Op'$ϥ..=,&bwLsV ݳpQJяO0ϙ~!9k/t~J:8AK_*tEġE恐mA!L#[T8HGu$!7dT2(A2X=ĈI8$!B8fP IQ 9¿LIEx*]ֺXw6cgՊ>ljt*9!( מFOk1:zԉb==g)h!AL#Al%fa9d Dd5"ҲUeJurR/Z٦!kOiMVj$ChgBc52X< 2!.AQɀ9d.) bTmAL&ʌdXGhk1glyf䬯czw+p}7ħî[͂7'yhV3PٛDn?GscÕ6PMF0&3n5#)RBU!2H/ |52J1hښ"uI{R^b']H)֝ 8g"KMeמ~I,S'F3h0KM4;@Wꕣ۩ۊu [UBSZѤrʊp,᪦dK!C!C(BK@(Y9q:$+yHIK!' SI)!GL#uZJajYF&P"\)B>$kqq,I V"d&1 -S;.SF}\Ͻ*~Kǐ*O4fa,y1Ɩf=]0w9iUF)_[o𮿔 U//aA,+W+dpFhfDF?Vpbg3),CqT9* t<]lV݌sN9kLݦd,7m w FVݯYjoiqW.$чJ=7]z%c;ɲ \\<9K"Q*ږC(9|JVUsdjɡhsr0v]箩`BI3Yj["gbGW(bKq09RA0LfWa Oq͍N6p/3PB$\\O 5n}}m)LI ÷0H#t{pAgl >,ެ\+ćq: tlwXaf64K|Piap%/>c`o>1d10ynaFԂ6,Wx7Uw8f k+؋ghw6o馟ըem:S+bv_Щ3Cow&7x/ΫLnF6#N>{U]*{evp3Zt;ԗQgfq{ˏ/OD|WBS9L`9]NEJ)I|,*ʖn"n[u.d}&VycK+!$HPMF^Kb%m%-%wjdp*T^kfE4 ]xsRDyop_.w6g:C.^$oH^Bh0#YV|pc:g49B S,:}H茝bs et*IEDe2q7滿-k+oiL0S-%>oֻzy^3^mNV 3 ۱W__2߭gJ/n?VpPlW'7>ǿ<`s^WӭTrwc*h?/W~?tUexu<[) فx8%Gc]c ݱ/;R1Sl*a.qttGG8p&DdJDFkcSZԣi[XwKGGM|nl~TMMG_VR8SRi'Tkg6I9( J9Flfx,͋NQi,|E?hłS^&z9y90I[{m|嶁u*F.3%o#qeZb !0Rg-ךR0 hCL$]T' |2 Ygd5HAzT,e\^ y ޗhCimw/M~n~_ &^ C_~ژUYXm!PKlJڥTOKT TB;GYR\!bHZWDT.j[9 GR`)~ybKq)޵B$/78gdxF_%S,˿TϐIsHjJ5~%ΰoSW5U>dae)up2ZmId\$/|:y04ͻ;_OL*;u CQtpP‚7bZ*gn=IOJ)rGD;nwT[HZIɭ0FIm &C. t$ȘyyuQ;H?:&(݅{p ;&L,h&Qg>JF}VO㯿)Re V-g0E#BG:ld6GqE"J4yIg!esIhA'a9P8KvW$bxtb5cLD7_}&ly7_=5O_et*W⒉&O4u<:y5C?drj;73ǵV^\mpKOPM4p-Q`"U i?p絹&H 0#y3O/CG>ZC yD""&א3U! %iPl )Y!b<ܡ _l 8fxZ VT벱>QLǫ)h">6Qp4~mYNcvJYYWE?JLv Aew8O  xvӷMj2{wd?2pCһ=6Intn 7MTJ SQ)R3һ0hQ!z0@L1 $泻zJ6mxhݰM3?١[[ԊV ܓ'Zz2h%?9zj#OG&ǵ>^}@z5YTc 95Nkumr L5Ym^|AUm*د ֤dt4MJ!q%:~ﶎ]nkkȌ憄1rxb:18qjȷUo7@bKZ_e "T]C8/dMR߆*jJX Kb H&wџ^޸3?'GkAb/ן.8Ǻ4k3=G;YMD#`ƕAgoȼhwK/>̾YՏdC|6&C˟gPMOf1@i6\78 p{geb] a:op:[/ǫ๩ifQN>٤ Ӻ{Ͽ YBLf0mXE{:-ϢHrIZ.#N{6,ߚu܏ [o4 SҰ)WjǫvIspw%Oq엚M*<=Ͱ'`C\-#X[ۆmszwAz.Sld#y.tH#pgՎseL  PΞ'rݕi># V\FVˇ}bs4/nz~hy,5op]4׉($U^\We9ԙ.r%n9F3 m+.q XEPz+sknnƛY?+?L/gk G_4_\6k.9PԆL.p8iI֞@l骭ލ&, {Tp6BO_&ׄRn5nu>ɶ^ #b$zsGQ|~}:AJ?g7jM,آQE4*Q&46{}__)?]sy{(-_^=3?~vҟnvGpJ?Y_5Qȵ]~DG/?_o Os@x>xV:$TXz)LM f}+54rg:x`Eˠ%Sh%k}o[MdS @uw֋x^Z5  4q0;c$ɟ F8Bdi2DEW,}0ZhWavL뙘h ʆf:NC'XVd:h|%D-*1 I3SPY\TnlM<De,<.C6,j>wqaK%8 A1TBێhF&JFbٍn4 K&橠v186jw vknIR5R1YGR'& ;k΁RT WPDRY2C hE{bMǕ)i/8|™XģFema<,6xXOǁ,"CmPfJH2r OA4*CHVLjXGdcM,F1(mE><(`) k#QRţ-i[Epv#eG'2,&%"+Ijp7"F陴bQ@AENH'.OIǩၩa;5=r,'WCh(=g~p8r57)S J/Y) |B*伇"0+ s 9bG sݟ0d ǐYWzo|^}eE؞e 'H28ze[_#>s0F͙'$bnwWkLH{ |e0Q;gi"A'kYN51 gT5*񨐐3" *{okn[P))\b IJ6 HJQk9i O]'99jbMCKLJ πjr-{sKj8 R톗(֋߂bĴĒ[LMW/6>IkE(Y#\`DUuavқ!^?O&9!6Đ7ZЀk1L!(O{ r%)d9TNB&ƅm`*Qf" DybΚś/ɗpDwlx[7nW- FhRC7E9L|N&+6$gLZc|! uW2S=`jrDq]@k=Eȇd1.XR5Bfۮot՜-Ϋ>6ůx7q m(#熱R0 kwup1=^}!Js:7\.j-l6 mѸhcA:8#4.@ J8^pc iPDƉP.-ȨC4NH484  pҜS;1NM~Q3z/+&jes}*1η~&OAxpTxA6&a0/:; d\{S13f;V>Ůѳ-ggؙ8;Ss9GVs_S1i s^u!,c,:}|; a.(~)!(^Iҫ0Im `)*'`R`ᤲ&}8~V1DyΗvjz;in/3fdXq @ըD,b w԰@p$'Y7{+S V Q c%q xD57HZAFQl_C5g;Iޖ9)&UKޢ8~s^c,n} W[EnyD桭rzXhF1!KQmjqX[FJJdvTRH'iYՏ*+ƿ"im]QvO\|u-\[vV';VcpQ†~4D*.'1 u lB󽥺S•F1'SգmO &׃ƽptz9i兟XU*/nEM FX_fgƵ%ԣ ("^ſLFCh9, 8KiU(E)9 ɯiiO#\1=VpQJyQt1b}ٶj`w>cj:opA ^W؃<Ռ[MJun{E#!7yۋY]gڋAX[ ,{p :o|RxiٔO d]r=tuxfxdTy_umؾ(S܌' 20%7!6U6fcq2Rv&{ I~<,p\lե7̣+y@U -޺Ƭ˅"HGpEKlªGN67 uo?='}Q]P)b"^! 2yH)Q!s)\?}{?t(Ӈw:y5iuHH`$zY3ōfP$13tRQH %Чˢ_g6dcK>l%o ?%Fu+h#C>KUȋpTZɧA &_Va ]G \>n~p(tĿ6(鴠(KМ;"IcH⦏R除r~vҙLedN4;&!Z*V~{툻2YpB K'Ep9!Fb8 )P9,(騋6hIsEbPX-Ea ed)9 =:B8L'CDo,AsX^#!yYCG%6 4q`wA1/|2*X3Ip˺<@Ŋ o]ņIOnFR.F*s,7;T!Vx`mJw77_?ǟǓq mv/bCw)TW% wϜVXm%P d=aZ[i^RQ"?]z{B'NzyXkDLB‚1D.({5R}~D5rcJ,3(XNsPlϓeMNJ%"Q(&t29<\1,o{Ah+3*wy' wrìٙn=:z!tDn8&  ,>e#8+h,:DAKުsW8p 0 YxksNJv~9!Ѧ:8Hڥ 5BA7a LO=l찐B~d( By<+ ?#]x4 "@y*HIZ8S9)Kpu)F8 "qKQbNpg *2傶4Ku֜=W2^ }[ ̵| M:W}h˻-Ӳn7U @0ʉ"Q,SiURCCD8z&Y4NzprϚí_ސu2#m{ "oW#Lkɓ0]Y ErEbT㊈ !*d:4{WY4(8=q!b(J%Sc2\wmI23^Տ2Of@ H`SbD2)ɖV_>DJrS" {/VW:`AB>`u"߲uLG.@ ,yýeT:Qg 'Zd{Van;U r Vjŕ<y/F7# d  q%;95ZJ.ÉL҅ܞLH ^6g-kHr@Kq]b p)m;xK gw-e?gE9OK**~?i.7W+׻M,e8,URONoQ+ΌOd-6N8#$y2۽Žv=#uWg_[G(: .s>y?=Ļ]r jߝbqi=0'e.u#wcn7ڈ1 G8n =ѿ4)製M'us#jx|̪$>[ +7FS[4jF5Ŀǽп<#q˿~ӏo??{.w~S/W)u2w4aڙIGO? ?Pn F;_32W4,h{q')E?YkF<ex 6]K ;tYY|~e)-ho !*cy-lx_o'~o`7n'i΄,@%E֒`Y t9k"'KF.:;sҴ!:ӈߊF$3dI$ HhJXDUhUB[8P #:[ -ID рE%8472Ȓ`3B p gFx'lsjq(.{Pa폷5\܉Z6žfKU6 !~|GRθKTs,<:hHy#Gee0o3q(̱ZK0Gea3ENxK!qEᙦcj)AXX;5lnvеm`1){0 7m2 n.]B. `e$n_noons8-}y5gҺ{Aw/oo|JohawsEw YٞƷ4]3$-z#B4ӟ+΋ݽfYD*:ls̨P,~Z`$锎OEb2k0Y#$ݡnvZCDc'ViWG>Qq^@AQPWCQ* ڕZiҲ(QdKslAQ8yrRN!ޱČIWQ|_k|k=nv7cBs9}/ѯT˨ ta@e҂[2 tp 2q:~1L]wr_ءA7LVGSŸukqDɕ:D 1B.J+Q9oH} dcp6֐5ɑK;89 v& 0hhZJ}LଋiLСdt$L1ꓥ^g|\ :`.6Kg2LQJimJ,k+W&n:ˤA > ՇUuBJD"Jy b3d `(m{C YT6퀍4t):2S0k%wጎY,բ՛{0,(k.M@c,~v',uLʅ5 })m$)rs)pjUM@]I).9Cu`3ݱ)s]U՗Û5AÛf^cr[6Qaj&nvKg9X~(2UPLHWl+F3niYɛuB~.?^~)ơ  !B ْ℺ɖwz1rnh P 8]ox`eށN1 AK'2j*y2t0J;“<P~|0jja1jZE`h.n0OyI?=p ݪ8}Ԧϣ]8ߜۄlϛQ4Iovum]:.߿Շre..]|IZk&ۿ*դq`jJDԍVw^4,WKjB_Cu}8)IyͤNϭW s- : *RM AĂ̮Bt_ ˼S8fBy ѹKslZ+1]B&Jޠʇ67| r@&@rRFs& }PB \XIN%m=byﳲW&a(e`HɱsWqN,E5u6B9c|s8Ȯ)y-:L !8!O1e9`L@#*ʨqnVټ3Jfi/^,|w<ߴ2nլGmg߼}c]__Bqwg|hWp>:J^Z=<`fdԦ,h&Kl5xp6 ^itR<%=^TAoFSy;_0'yus6|>{9826:pH;R i`6GŴŀ@&.r\h%= k0ihul1XͥʌP߮#m?$ oެZ>қ !pCiiZAol/t~D쬃 ~^c]68$ g+m#IؗYto/0` #HkTȖ5,"iIQ)ʀm̫""SA 'YۨN*ˆ dz;*z:$LDB @Ar#"y3ljhT>)\RLI')30t* ډ3K9RBPyWdc|`5 @7Ƃ(B)Zg:Y7E ?J M^f7' cR !x.AUWn~`a'/ ?{ڬbW ;Ĺ}GX pCa8Ѵ4\¤>O.LǓ&`Tli,F7^՛@BN st܄>>gO \g?O4_̟>rjף{$CeZty;=GV^N(윱<Ѳ<ѳI^R=BA>j_zR]54y%gm\{-VZrclz nW+Yjm׍ A^uprn0 {85trM4z8ℬ~@NMdg 7g\~s>!p62(2&]TvփVA^hPɎ ZO0\Mg&ƼXMY-ŀMƕ5>O=kyKڞ|βsT(+P2HJSbtwv"9k^P˲o@<~hpt8 eGjF3br$17i.ș.%ʺ;2B/S`HүUumlɁ.Pxaq˪\oG[{RLVPv$ 1 K+seeBltfI(HCå/Xo\ ^0{`b?!GΥ=ԬEHN%'5Q^h4'TБ!Jɑs++e8}Y6w=CoFF'(\]IsX G'SA.i|쒢x#U"e&)aDZ1`<)M "Ӓ ⪿{3jY2mSG-nxS¾dzIJJ%Ttr5,F-iJ X%8RT$X&U=$n#!tLEsZQPXmy4hTpF4C{~6E>fl^ʖ>L{|;@IcZ~r(3C}]=ܦ԰|\_JN`Kg뢶H1gԠ22QB9agW䬯甇3Wp7g mxRV Ӝ (u ӵ~5Ӎ7!ealhڤjw:vbFpXοM5dK?,[7h{w>ڰqV\, sL`gFMxd;͑П{kZ͛glo'?7Ϧ>[ZLgmVt4/{ޮHF̦M 7ZmF?BFu$!m>!ìGdq#p.|<,7sp9Z< ӝXF].mԶ mXHw $4Yj_-s[uN Fǽ0@rï_Pf|/~pJۺ(m4AX fe2~2CξYz6lg ih{q%Hʄ J X6Ӊ ڲ7Nr@d24%{q꜠FR:Igyuj&ܠntus*fM'7c޼]WiroQk{J6=5uc%iT$91ߍQE1$ q&W:(jfNcU)@"AtKl?5Tv18m;!؝ XըZCrZgA sHpRKYA8j&Ԥi*B%(,MgD|b]-lŨX EN.x*(Jr" .nQdKs$L'C.<,F"̤",,W}SO9#Jz'IlqeF8=y?>VG~bn9 U _qOLbTǜP&Q7Dž8:!7bk.F<6j? G ٙx-$P,K9Q+O ]>3v]{^‡w]Nrq Me WS4oHWn֟[Wֻ{:]S)bwfZv~mg)=MxG˧o|t\A<[K Kƕ)<%:'5Jlwo7њ ՘G>九@M3'/Zt`a7%~:>6J5ɹ '4LOMxL1[V8jWݷ.}~ߺƸؽ@SY.;ˡD s)7Wej|nyٓΜC)uҩ g6Kw9u;ecs;aBOW9^bDSG!&qލ?`ۥL7k)Q'ˏkyYƉ|񻜅Hf&Ld}B2%,ݱI;Lh]aMK-vݘXKQΥhH+i+_u@>ڮ#>mw\硫N;2YD~|jxJ[m).Q)k$QHT(z/4eM9%YMҒZm4E=hަ6̵s{_˳<''q};Ɋ5y]SO,9͋+n_ퟞ|Xc۸TGk'2q2ق :Z| I|v[mEȚkױ9Ft8C/PЌj4uASq} g*Xn*b-;US5C (iv+2KNָ(}/;{{%sHo,8d}ɛ}nK:]޲*ْ_]*k$9_}w0X7 ?aGwiýmU|a}}^]T!_.m7|f5Dv]̧Gmۦ͕-(SL q^7O4쯟4LpXLp>E Jq97Ϲ'G'{o M߿H5EmkqD=kxQJ7^u z/΋Kyս&|_]u=b0\I:i XzA6G'fm\_UP\朏~/#p/ <0U֐gJfO5ѫCsϮv[J{_ie{=%wkrrQ:L;|>d6Z $s keЙxV xХc](!ph܉LgG'\$B?vю?b1?_L=ֲa>9z|s#Gd2z?.MԮ]Nρ{oݚ=}Pu7qMUfٰ|F>Uʛsz-3y<pe={gOU5_g39\/&.c䘙|ɭ:=:7+WqH뫼"%bފquf?l.{yo#k#ߞZ(w6/Tkck.D-`7w5D/%UÞ,'gVN=aXsH.k&75`Bb;suz7֚jSE}?%9g[v`\;UZk)Ek\$Y%GZ6<@.Go߾;ޙKKl5`eJ,i3DhyvẌ́:.̶XD38f" k#Ki1q!P`-m̔yZiSvƌ `{ahra:sGWa 2.Mń 1*Z{ԃRטw&>#FcSH(pׇdɛvv|]m!jGK{)#3P2ee9S*8v>:ɸs*hIwYȪbiE[8gJIQ[ۦMd:cK n:rMqI<\bu&HKv!.qµ"Xh],ԠF bvj* sU=3QdD묱XԳPB8EѮBPKͶn' L#ˌwhBJ~$KK!<ݫJ 6p AbKKN|g%âaDIi<XeR+a x<{[:8VEh#IJ3`\p-! rŸI. *AeTk8(-jXkٌSPvPP;wVHChWA:є9c{)%p F.7!\"K`l`) ΢HēdӒG1Z((57cv4*,SLdXe0/}KL+yfȸT ++C@zah4vc.59 E nB?%c 5 Qࠤ`gLJA42~A.[:*O p!VZ!hTcL;3F@(GM0 !-Y2(l Ύj@!N)Cy$`,tj{QRR}JƆyŜB¿I%xKcE,Ed$wKIB dHC2쒋-Q)P&Z"AD:3JtK1~nc3uQQR jtB{as7ΝO߮WkV ^lR ]@1I|8)x@X8piL N`˖8Qɱ'r+*GRLe1' &yc9@|ѢBBM"yjVPiD.iPy]AA>O1;j 3*]7hY= %*  p_v zn}G-Am@CK7 UeS Uߙ}^k c¶AMRBNEmtrs{9'yQ'K3!bc2,C ʘ"zإ Ϸz΁6D2$|cG#\ϾsIW3S!vo0]J,qu m1񢜑c,#hm/qt 2e#븶=mWü6(HT%kYJRP?zP*ZGhw""aW]qpqQcݨ|,,%bl%'J kc!]ԳpgFMj$,ʬ ޶3 jnyLUӀ{9AEkLmda@?棷JPafL qmGiѬ /rs#]Nzд=?:l+C(@9ԍ;;bf]/- HaMÿn5%3,zjqcUc8W5U@,/`{.J3G9c,Dt&%\Hw_Pz m ݕQ'6:San0zC @c) T [QWXjJ[Nnɬ#wdEyc(qyP87dQH0SnQCB=^`-\3:ɮTB?&>(>ps> 31XV=Ԭ;5hqVٷ`d* R,Yx:T쵻?ٻ6ndWXz9U~HoUlr|PNPPC<=OmUMuJ#w ui+U>Pq$ٓBjl?e(a,C_Ѹ?jhhH/ܸHXJ֬[,ww@]aLR0MR] '1g%Y2SWW׽<>Yxtԇfr Z}]s݊\Vo)-*Ж?(z4Hs Tʹ2 8~Pi%xs|`Çj,ueVTx.%]PqV"[40kkø\fU" VAN^,Qboil4xX%8DHQcFdTk8;bdJ#:*b 3,4t@_B TF6O^J`.qgifj:Tg3Y:{߼|TEHpW;2Wq_hHwEK͊R{WrptpM}$ ArBOLmHUʀ6mj7:hW/d_0~E,A=x ;zo};}w*ġQӽobyOKSU{{=sS+v&Mߎ)+v.ZWZ(rfg:O3e]AH{+#fK kZ`-yWF ip*%U #; UnLYЙa(wAքP\ٛPV#K6J}(i0 fԂWZNc(d!N-v匹.vU+-l٩晕MZA\`fO|-7qy!l(JC*OhPdjy6Z<2:zS\Ou'8 ݠ%ś \o`Ίg;6d`pө A~[wӯugim~.Sr2.mMSׅVS&jTҎ71;c-$/*\Xi" &kK&!Tla_8EORfO,Nы2.Y3gXR~y𾘂2?3 yBS涋eǚWφC?/}7Q? ;[mPW5𳛞,9 ?J{.4֪Lq>y;2t&Gm}tyDž0ګߘ3O+((h4\p6a%(Z .Re)]|jVju uO\і0{:7)t6z,̜mA00 coW4}߅MJnyZZ 40Ufb,A@VgD&+ ӯ7vFmG =;dYC@V-n7@8 2 un/uGnVKMQ6ʨ|̝.8+Ĺ6z](Ow;CG0_uq{!GĹ'_NO¹шEЄF&a'e*N&eeM"oiљZɹN;4Ĺو=Eu_1X|֎W ϯl EצCTqCz1fiB]&vb^VITJHa"U]/yPuO-P{ӓٰA9zS] 2ޭdӝ\\r~N&!b[l,| tpn(_y;Ÿ.]./2E-|n.8Z=QyLDTي3(Dxp+/@ .zOAM; jf`v߶!̓0LYn;aⰞ刁ӏ'켃%:CAJ8+e1 -pbjע;E w22!dtSSpFN[<&> Asd6S):\LxꂇޡUyVN>J cO%"w6MPenE `Zvu L12KAG5ն)%RmT%(וUZTڒ`%1$H~j;?4;M$$d2Ÿ `۞Ifb0ܨ,Cp\,0+jNv'Vn[A) zzz;le=2?/+>M`fHC`4[ɬ1gA9B sI2N-;ĹقY_OFi^A.}=Q߰Yq}LD ,(jyd9EHyJ\iC%4wXr].L0mUBYd\啐к9eX^Xn0"G^<{C*Xf@G9)W{ϖ9P .U.f:^]\!L->$?]`%C p+n $~<͝!ĵڭfu=:RF ?+7~IJ +ሯ"0V!>@"A~!H+=w*xjpBpAe6Z)u\huTeǃdJy$a>C' Wp&mlxUfo8f~UC_zv'A>}.iI k4TyufR)wGqe 9^k'_]"W"9^y\(a!qYi\bN8'u2)yLv,_@y JsN-Wy'KrZgI!X8e/Ox^i*Ep hԦ,L6_ <*x?ۂ߯d)mMܼh"w̳v7Pk-O2Ψ%+F?ŏ/] 0W1]e,c,yBb]v<f5!=ɇMu0:.]_[;j5Ӫޝ9XwyJHFxi;MTs3+@pnr::F$&ӓcAyL9$ As8c}F峲`U.ǜzb =OWU U]8s՗;gRnlٺ@Qm BOs4>YMB"c;_.ѯ1֋bn? $ԭ@Qב], , !;K]8bmU.Nl |'ZƳ\r6L`-)x\(Ywp͎enyoҧ>)d;/[zpZ\b!RmtЬ$A*W8zo-vN_tV?IKV7'qzrt9GZG+SQģT*XeҚin.'4(n<78z]3F~к\~0M3MMfjdе-wpl wo Lwcx}ߴd%J;>6IZk+Z<$sٚa] '^yRlɚ w:q~hWeϾ3!˷ |[/zg^:/9v'n%Ϥ|sLY)t݇g|-,ֶ=Uf.EowŚ$_A#aܴb,}2Y*l?||tEl|F@!t]b뒘K/_mr}رv+R.O=tc_*c0O<Ϩ/ d>^o|.m]4qɿF O9Ђ \M{l \OilHml+vfxڡذ΋esxi-"aa ԖY*2q;mk,`=SS7Uz>o4= oݙn\X&[.p_$]a%!Mq;ą;gz8_ T/&{`f .$)R!Ț9dS&)e.VUW׹w0r>>⣹}](&#.Md.-R(Jg+5KEA.pΡ`xWY9fES G驄Ĺ&k b!q:r"V)9p:"S }vXGdf_O7oQ EKa)uNK -pO]x%EQĝfAO` Q";m@\wo|BﬕGճjҺ{둌.t7xf+P}c9\MR*r(QɪlԒĀXZ-EaNeŒ(n#!tLEx9">&MQk{ח1-ux}wA\oS*V|w@5t5CP&ex?Bo q+>M~n>8\H}D{8=!KHǜQ:RZ$R.(MxB!kxUcgLʃaVȦqį[y4U#2[8;' *< .-}?9"770t}ϟ;%!ōzzm9Z|Y, 1! \VSn6 ?/0^۠\swV Ӝ (u ZB-\aPygU;r4W!u:7ťrpe_-8ыKeXRSay<!Ĥ>Z` mգM+(B`/3-[ٽv uqMu|yuSBdpf5 i8Nbm]5 n/WW܍W;KAHHR7$.a8fyaa7{ȑQidE&2hDžGyR:-`n:YJYП9,>sJy\4oY$j=mbm*"":v)zf=IYc͛x&# @j 7L&\misɢ 풌T;oe `9ˣ骫^ϝƐ~4a(|%02uQ.B ƒ(|S-HLƞ:DcSX;Imc7:Blm?F z%!&3em!rs_ gUB-K10!tpbw7[O,sI5{ءmȱ~O,溰5GҴ}MjOJRލ ެ ?Zв`C R% < g^z,݂âw 3L#rR < @ԩ(%S^hg-&$*q1zbe^G4? 1P ༷T,NH]rڟkZ>jm:v...^y+'MAO.vgeKRomi IlPAmYN@ER@d2?K?Ep, cɥf\9'2zLMg?4o)0-=+Q٭퓧^x> yv3 Ѭy >8F58$+@%"0bR8ЙG)E-ܺ( "JOѮ. O!{){xEtP+5Q'툱 ª M~2k&fWvk+\\`wf^KcQ@Uz5$CHҠLFZe1v#?䈂9bw3_'T 'X}EKLjf$Sg*}ζ)c J AIiʣ,e>PTs0!%̈́NSrN9bk_G8+%_gkXZ拴=_ܝBe(P)=!.XՅ-ڏ xҖ{8;Ň]ᇭ펮#Kr= =!}nfsdA ; ـ' Ez<ף (aя,c.b%—(.PDKFud}039.Q`GK0G$O$jFNx @mDrfIQb(o!bHsl!tIl>B]^:Y4<[غzܺ\=.xU=%~jVb[ûݵ|xKx@k-wd,|Ȟ?r<8:XOó|w$DXm릇&^-)fyAS&{e#H&li>NOnDzS(POJO(-W1}Ie>|C /wb $HN AK*5N'P,h 0yM""h \4izbybxI|u@\q $jj}y^ [;4lzM27ƒ%LI۫]?]|s1\mX+~רEBU6^NɏϼXMݍiy5&v's>fx՘@rcn=]l6ܾXm?bt\/qWo^/֏.Cr7/"Pٙzvp\sNU?ՑivC0$ЙI!q=F١tPje`z0#F6WoCQUAZWX=9aCIovO}YO+"Լ^2qy^37߱^*ΚlU/.6 = R!+ss2gGHb\S*G/6^ }׷,8{ך|cPd뀂[w[^l{O6j٥^?^,~9>9-F$z6N2븷 (S\:{ܮohE`-<y!_-;{ƝǓau< Y%M:Y cwWX_g1_6qu+{w<}aoq)~)=`o|3%)Tfhf jHʔE u&L{}'2dYZH[ ӜvФ4IxPQɔ Fk'D<gOb15ZErY`u^ '0^樂 "x^\?n,U6-mJ`ѯT%HH/%i D4pPZil,@@Hv'UP:o3V7_&8M23u&߿XeUU䤡z2qa:IB1LDaY\a:/ ruQχ>}OWwi[2f=f3B$hKhf^#xuPKq@82*}t,d:Prt9|h2@]<"Y7mq(hVcQB'!$RFVWe2N隣U8LtaXry@SQH\t辤"$Qkusx$ ې, xSƢs*Yȩ ՏDE}a<>QI9mE5Yr&e5Dw <}X O힆>GUD|ҷK*a[е2+QwPtv9(P"Q E݅Z˅"bjI~^ }1SPA~XRr,1v_wK]`fՌ&"Xj5YҵD5^}3C~S:-C!UO"(άAbD"~4~F`?A.jأvGycQ5EOCYeI6wU&BPDES6Pk.'=]DP5VY]xfP R)Ԥފ{)pM iorB jT_jkٳ&!Sj1=hm6a󼳛\:덖1xyZw1$ԵhD0u 4tLCGs ncJ`aاn>|w9Y:Zm*5f՟Fj98)E.Ę43$/GÄi*3B*=n(!/k CC*樇2tyD0Cks"Y'%\`LՃ%L(H H*,m􊠔0֠7bhlAT(g7lE^QH"VR2P$UAU;Y7H. U@ SBT/2j#5JuT/FB,&z"Ai,:WQti#ʅ0ATj hNLUY3fn4Hk֬U%(쥪Iu2 EIp՜ھÿ%=kDo)T`jL1JH؝4z XTic NRU02Q3 _B2\HO4%7aBD5>d8N֞ JOPCXe(TtqHBU4:㽱l?g gq_z;hŵ5]w2.{^o_vq |;|[GnxU0\`2Z_W} )={ Vc\+b+b+b+b+b+b+b+b+b+b+b+bWDAp<j<{I!KW^kW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1py+9i :p+G`Q \+/W \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1p9+/4W`z(;pխ"p%0+b+b+b+b+b+b+b+b+b+b+b+bjV6̽@qNb p ;ZVka4>_l;s//O}ZbYu>sH<-_,YqqZY>8}ut_/Pb}<}8oтon۰k.SʆZ+4wg^,Vۦkt7YQ't2@o^U_aFEuPTr4U=ʰrHF(Cj;*GW-˾z"ڗNdH|B=KseCӖLͨabC[2QN褸vLS* x>Xn"G7n_/.:\-wK|:aWvc2(t5xsS !*ȩP2<|͗/yzAtt=7F%?s,1tjyRs}iXǃD4a?0KSq*dP ?\/_ xcV?h8?;ih7<籽kcLW URCu [#cW섴GyQ~ r(ȜF"}hQP yKd5nd# !gzH_!{> Ì 0{@ #OcY6}#EHxxYb%"2L~dw1jG%}}ړ6h*[ڼjuII8O*yNIdxv̓ȔP}3!$QyK7#_oEnO#dd"mq080zGu |~<%ac3ެMymL.<8'ys$I!}s;3g Om8_Vv09do⚿V5`r)r8'vG,;GpԜh} 9Bb&f^(̝o 76g>FӘo峛_.jrDaxլUu-)L^A˭=/.ނ&Bh ?ϤحSL8I^~RQ\D :ER3He!yTWk9Հ ɴKaK6 !2ȰD8դhUӨ ] M2ɮ<ޮ88; cٗ0 USU5R_F dE#`ex>hJ `dƈY2aDv*=K,r ˮ+:l:DZQl*䩎LjɥWK2jϽ>w!o.BƟ} ћb:Vyb<#=}{:9'KbPdI u[L T&#Fki(V 76L@]eսfNQ߭Jis'-`S.fZEe;TUPQ[ZB+S|cHSS;|^JU{og*(]6RgMXW1qҍMl*Uk׍깅xx ?tm{x0oy[Ro^EoN|f?jMбZʄ2ZZyMD0hxTQ͍QSa(FV;|tY89_9^C)ϟz۶V C 7a>& Ryj@mFEa.c' ժ-ih>GȨp΅*aeܢDA0AxsjCbMk6pv6ي0y[:yWs+`7%M o*h! &cD!+RcS@XSh) xԇcNtw̲\b"`mAְba9*> nyo@~}7JZ_t:+<°3v |S1:o"r~mlogy8zVTbҷỏ4 lAP6ۨ]1m=J[Hu!H5dj7h+[`h~2U)b1F13ɬ`f$ŤRRYn:慧ZI؅TUM4J[ 2qL:bרdqmȩxZ2hXڼ8)[j>og ܧ V_<[tt!t*$ӣx'%L26&\39-2Mlf%'7G$b̐C"ZjXf pm(^K/:N i/[Bo C픶N<ˁ2;R8̓<`ֻw{~[BA NJJ:@lٔf8.h a8G0&awqkFAê "wK%Yb0$AHI_6wh'|=g/IxN^A?N:i޹rkN-ʑﺉXv}]ʼ2 T3VjO =DA}L4wUsOq(I٬-{v o ͗Qܶhv:G33|ApQsU>cXEG 'X>v yvC|HVӉIW?-UD8!Ӕ8c3C?H~&o0[`Zb/.Qdz׺\7zy+Vieäc}·ٿΊaQDC/VXgOk~N5ɿZ^ؒ"/cQNs.<_7~qşj=&&{4F_1^A~\%viͧה0Rᬞ[FT*j&Ca6ƓS) 6ke 3ߟlYn^Ŀ 1LT^tEs|aӬqާA4ah5kڨg6?h f(B <'@ŗ<:"іHxom.b 5tYw xN_P?OGA$Z[U@(&w~DP% RX *jdGO&FL;f^@mgQdQ6L)nK#18 Ō93˭"Ă2UHZºRN ZGK!@3bО<# baJۢ,ZKqo|h}3 iSJ9w9yR0l--ptzzS|}ygCugtL 3), 1A"A bbP?r,߅Amɛuʈ(FQ @]_yaye4Pp!J*6!`B ǨU1kQRk^jF5lkl8{͙һi$}+?s^c_?{uoV_ӣ]je#:.J`zJ) ~BR>h74%'AY;xbxy T8l؞$#)#IEdQpUafXTq8"` (O |ɁW1C ( 0 "Zטu!EpLh$2VE +gZ=%UR}IPT~ ;E3)Vqdj ܇ef阻SSYsc.0!#l Hu >ʤd9Aj)|oڪn.χ., Ԫ;AE;?e$;dJ0팞!II1)%fƅ%2(aD?>#bhb-}ЏG}2e eSD^ J1uw&Ã|˧ @g 5À$\tHqYƍb+,H}*E~eFq2:_c:] oG+!k{cwa;og}J\S$ál+WKxHJ=@3>:KP~ dA2^%XɿeIttt8:?*aq)+]d6 _YqGH.&`cqeN㯋d^'xWy_t1:?x,x00j5̹n^[gk;g Rz1Fg7]n3Ҟeݐ(ybyCe`Zt{BW>;Ѫn5Xdz]6W<§V&<3>?t9?E+|pѯrlڨ+@Սuzug@~yׯt7^ &_^:hټLfMB/όҏů_ʥkQlNV8uNG=_ɯuꐺp*ܸ #06_Pc׺ůZ657*A59o/mrKb Mu>n@d+X^leq0Wĭ$ݫ;ד`|ND@` T@yE#XQfMY6;ejk68LK]4:x6F+%H%H#Y>i83qM}+H=s\I_Ni7FO?Nޝ'0iUXՆwE aPs/gS?s-?_hkb_i$ɳ5b<%x(7؎71yߞIJKs tv8zю|LO0>RmywW]qÔj9d EU2n}*OAn*kԇњ-%UuP?!1^tq_tjA>&N?_DT9y3z[D1)'ɍ;E1 J9Ij g$y 99~7Z]XN4BRE}SSz ɞP0vY4a#A2?9)iJ4=XedO${ZBv@Ƨf (߻e^;h/pضH두 7X JStsA:VUs(+hЮ KPnehdxK-F,6ݭX4$XV.`(6"ehMrM[H9jrА!?BJ#QHYu2LwZ D佖豉hj aTcЬRdREC^qI=g 8D%&Eij,Na(2j`Hj` <* ?s~%gPro066NK0U6>ܕ:S /%&49umɵsK`5!%]Nc)4(FR S’HH^ϵs0t$Ah1BQ(G H8=`W12#XPg&jZ2669Ҙ,l2* u+ ,(mt񒢭\]Vnԭ/~AݻjKl_Tr.j(1 L2Jxc[bbYhd I3ZBS@B:`^e) Ӂi]Iհnl8%+R1["ؕ8\iհRhX:ZH=Kj8 SfSnD.qVq@Cdhю Ƨd8C…{o4aypìԧ(aLJ$wE"6Fw 5+j%b+W(CEi-aATkyZʼnP4+ zmt7eePHq)\S/IXo84(@&LR#1s`ᬗ_lXc%[gcTrWH\\P 85A8"hH.8A$F"؁V.nC.I0DZ8ڡ0;ؖ4dF{-‌TФJ ?3 X{`g±!rg .6̱0g凅̈́]욦ǧM7g\9>|V7ZM٬`ńnl>K:?}]_nWit͚:z5se _N{7jy47oR>s_-y:ߔʞ5V7$ݜS"vWcqxESId]ɤ%`|TZY 5FIvJ"$K`%V0"XD&+=5 :CDǘQxayz˩4mnRm!}ʝ)i'K ) I ٌA6 *ƨrR$>dT!$U1ՑFywۨB562H[~7k*p)R;13X NG b`BIF)b,`3I{oEezNwؙ^oCu骺NjV0I:#J-]q]O^?6Ad^]&I# 0\HjGbǢ _cxhh.+$L>z[?ʶ,Z=o$=h1N^9y[Jq] {UGA훢v=] sqp|pfn50%P玟Q $oo^XJ{Pb G.-Lwcu?^TIv?9L0+GA7ދz-6AfP&IﵳZvgrKꨃ dRKJ1Szfh ˼"{)Bx|qNO6U}8we@SOUe^ &sx_^|&i6^1 +$c1֎OeL!|N]Orޛ{Gԏ{.ƞjҌ <+ `Ya s-sd y*K5{Dis0|=kZoWWOgiKv%;ז\FUފ/'VsQޭcH5HU%W9v@ &6 Gxi2삤X>Xj㝮>xtMXRP(v!Fc5Q@iM&(Cu T3T CGkݶ #gE-dn H>1I@y3I0\p&qYL2_ccXj-HϕM9iy]yor* lp?~{ody+}J)P8-$JÐawuOտ 3 eHy OAsA}\iB&P 1aA8YRQ@i+'"l OCBRuF,Rg2TG%e3p"7Dd%0l\FPgnL',}@;̛{ZVX -M0k?!SG a" 33p`XTt:( bVnӞy;az{!z4e`# 5౼ +v#þ'o VSw?b襣4фL`s3h&upu=2 Y2>lCՋzNsнk+ފSCLSQL$Spo-UށY`f m(uj/m*؞H>HԔ0` Xy$R Tݱu9{έq9_JL?#{ZEn5} kmR;}.#V|va|F7L#Aj"xacP^"4!z O 4FD1:mN6)4$Xpav*pZ W8*T@Z' k$Iΰ[zXlJO^]TqBV( zz*d:YP= t"0Zм%(^7,xřVK&yxDI5=:|OӢע]_ 7z4z9/.LҪol1!Oʉ!8{oY(2Yb0] 5BV֐[%!b@|X$bVp#ml>n/~IrQnGsOݾԷ4n1)Z&(H\QM|6'"gkTgd7+A~9I`&QR+G`QD1Ҁ }ܥyqXyB <6:Ou!۫WO}mYdxsDrXR96ϒ$A "YE )ĈIg؍K+Ӽ=_F%&zqyq<3r / %#J uכMVͦ.ni>ݺ&llr/X0&u"PBIX`kƒBh`5TOI;IP;c-<:_[p!w"URM̮Y:R")9kN*"ˊ 3âҌ;p0#Ay V<⾖ 4LE uCʕ8cH:,dHVE +gZ=_OSj$մ#I Vb9x"`'b"hGko}HZ8=#uHw;5?Nħ(@~,Exi,_cF%QkFvF3XGw''~p:y?s3.KDiiR#8˻ۨpXURS|%]%mh3|%CJ7&]DĈ{OoOub Z?+]u 5]S.G]O>.& hٯ;Q`Cl%Vmb6]߸$ډ(eh*p$c^H4VʌoMᝐz?pK͸JI屳6RF #HFΌE\Ӡao$O=s\ Aiw&ww}֠ڛfU4;L&ugf-;J2b27ۜY[XN] dCt,'s:N|<ʁnFVG"FI""UJf$x!δV$ZN>a䐤XIE4hc}J!LE N<Ȇ 댜6}B6zbOڒ b Q(V**b`*) }7}@QMZf<dhMq,$.;|!GO!}UYL⟯b6/Fׯ4+!2>OQlţWG[mˎq(9:ʎkKpY(k@= <ճfn _+sVt]IZ>B/mY~\ۊB`|r*9.7XB9Xqy5e(2r£~coN% Wv.Df[{Mܚ›- 8`rvABˮֶ^"܊IKd?ӌXD*t:I%!/p)a>Fr7$ ¡+CwIN^v mWF@ɷ%H()VP=ؑZ H((cmCݥ_8WqNXpoJ'`%:7ehj'Ia9-і(4;M.őp.uXEn%V`>UK4"JFcD4Vv;ա&<#qH+&2*ra哊Z,r.xccPhkܷ wX[HFm$ bnGtS <4d.a퟇t9eL 1{{K%`u<u;?ޒ*&Ζn&{*zQ%uW ISɥ`.GY[nqZZ#pE5 -Y B H6D ϒO hXrHU^mHaE-&fE<K8Wv}t/Gu͓^so0_'jWŽo}x. Vv\xS[O]oM=ʛ1ZY$49MYbݽ m 'l,C IV2*@򈔤E@┡AdԐr`RNGPTq|)\B#U2ge,Ub3㾶vY…D&(:Focn07_A݆{ <ϩB2Q(M@KDqQgu '"QrÓאɞC3ʀ+JftBnGML(J!&)lb(Xbc_v`צKct Tj8tT&q#{ G%lʍ$vI2dDT #QQͣ 4;ME [bl"G3'_gYr_ E.ӁPA.VƨSN{B%>J [,-Rq.bc_=4P}5W$kV%~|o7o(ڠ-'~0 7s} 9HfTC.!|a3sG3xH0Q9΂IH~NxgH Z@rE\Fy ^("9Kr~A҂bS恨-izc;BNLyvX~hnBg8JS5UoP_4f~طm ¼bӱ.(ա&x-;:QCGy&qr>HclUo!aV.lpN5 "dȀ2WɉCDw0oZe|I0 &bU3Qf&Fz'O- bRBud.5} vLI=K*jUm &[9OZbrNgd"} |#|w`ˮy \+Z3DMhG#s9?!eVygY[غ}ܺ^=AE {~l;Ė5(l{ƛ;f'[r|k][PY:Ĵ;vkR,MMrwئ1ygg<$l}E7ihenW43xՍRe+G:B-5+R{ \AB^坳]يblܣ+[^x$ǂ)b&1uJ'a%p'bpLE} BhwA $.>Zs|jr~$x^T{PZ;NєH/ ŵrQU鍡jLyEFHB&`d).&*Ⱦ&ڕZkI"91T8jNQcLD'$)kHI).e3۲.Ġ)jk\yD#x^H0Q3–[G U)I2/oH'klK@d oukל!;ֹg:9՜s uvH5ѶNERmT-(׵UZڒ`Wمua|^olG %K 7&$DMbgBu3 ÍǣI5So|EsPJpB&Ct#₩DUQtGj@s€8[2w&ÏW{CX\ WѸw9c[l]wjw19Y`- ;6bWZ!0dV LOƝdc2(& *He(Kق2apVEK!W™he@~d'IQAR Sg߉=q5idqWұ.nN1X+:fD̒YDqfٚb4f2:n&M )lp$_KCi-1N4 jZur7R=vޓ=JxOBR $8$Y" N"ug9BPڮd [&bEG\(B>$kq zNHHՔN!d&ޖΠPJtf)=RMGs`<7<,^O!OWegd0VqڈƖv= `ALWrciyz1\+VJP3_j4(_t-5XS‰ =b# +fx HRh:n 73$pE?hF:Y4g3%CT{*E +٬ տh&K5ꟿ2gX”Fre=O!01^G]e2p-CޘǝͥyK$yv,v^˲ NEXVe@qít^]A(1$|: =ڂHuFu0ODae00h1b\m2~>IB-SAd9Ƅ(іa#a&qE"TIR3*uDbTQ Ř5JƲb\J8.p"$I[UI$% 5Ƅļô҄Ҿbllhz)oe&eS$ǏMjnC@Z7qZ_v, 1Se9W y13qaոh7|_/pTӗNU-9"\ypy>+&/Fz!R֧^>~;LBhi=O0,< .v"z2g[^uڻM%2@):ᤩ Gd#^>`aJ&C B̄Oḙ ԧc|tѯZݯ\75Թ1;||}30S޵-PҾZ@H0ovgB=^ȎYvL,.иe//,wFFT= ]s߿:j&ȘΏbֳ{ĆY hRHn. Y/k7Tﲤ7I\W1+Ēm6b۬6d+-,5: Kvd.XpSguBCRO[Ǘ^WMtWŤ5yWc=44ξGdMNd1. qC's ǐ"xaaUu2XO.7O.M 8tTgmҝg+4FF/rDҚDF;., Ii0yWsd9yR.~ K > _w=nHzwpl\l,E9beެfzl!q"wZU GM5 H{MuUB[22dǗ4,nmk 5aS8 < JսN{G1R` dpi@Ò`j~IX~I+b6E[-%aɇDFEJgN666Npjeٛf ڪz[Ny{0wn“I]8YJ +!wF2JxװAƨHBcA" ),NK0l*6Дy`DuB(nlʼnt}w{etq]7߫W6ڲw}^M_\nxt ZonhW~:ٞ !-#:ZY[Y>^t|=nž"}Q%X .tLn]i3J e}r\=zrB*hIBI>S] AOP',.X!w?Rb)֐£PJF s. I 0YǾT3q muCj1E74z=;lҞ=Š'P2"GQLN)RUc e&yB.H,e !)(;l$N*tNE9! 8 6+0,-21֪n֥ԥ}z} N{{[o9;Nuz2W %?=Z 5ʻ1#IcVd|Aޢ$)˒k+uԁ@VBHe@3Z(gBFJ&$3)zY%o\KD:-XO`t\[{fJ3_L3 _/\HT" ާ'lk̓4t~t}7${f U4(K);(꣬ 1k/.BŲ+< E|(UM(ʺd5Z s;IR!o] 8=6^_%9L;k6w*(j\RVlzRHeEv cd8#6 l۸u1n2#C̢*}M.S>Yn3,mq<6g?IJuC@mG<;#~-Ow0Ge )(BKptT QpiE1 ɠuq1ZA @e%U?g8R,8U|J̤v*:ЩVjX8=⇳¯/Nq )LK__IϱI~f0דh6,$i`L}5.0~s~at_j'i\(X = w^!tRIs1RBl8*!* f|;k]^|^qϬui nW~v]l:98dS`L-LxA}=/ݨ.\֞n*M/Jkߺ޵֦c1O6ꎯ^M~{ΫɎ.28ߐy~nOpq+zgxx}t~4|ǍgP%X~[ʛ,TsK-|Mzs(O^$aN$pY0hvӉV-!Q c&(:l&&c8h&q[B06kiWNDHGU|0EiKpX%b2mb o:~+4,D c "Z#Ǡp(2Bֹ W͠)Rp1HUIJ\A. =P lI.#1f| n,D'dQ‹GGJRR$ I(` 9xQ(9K'c9 j|ۿ;OyDt`p.cSN3 Y;{ʅ;N.υ +'TO "LѪLl6.Z ^jB־S/nWȳ.:dl |FÐ `B_%@LdLۨ F{HZ-f"X?*V^1'2x jWW_t7tggkv'e6 c.wP ;J0W>uf@VȚY3 [nRI(Fd D &@ղ$X B6bn DCok!]H\"\d), c> 2F1m&Ξk c_q@ֈnޟFo/k;4́ܐ}hZR˳h:H%PhpńbA#9V0rY$C־Iٔv%%jFчSҰ!)iMl#3!&؈ۈkrh=7jK7\hQߐ|uKM^~Ue²o9q@c?"貄P`Ec W:)uJFQs%ꑘ*|) z^8nc>./h)S f3]ڰCqN^mYl =#*m#8*s.$` &ZE’Vf )B{Wנ㓩=Hmn' BF:+WiX~gEs=MFu&ʔs9&mJ١FHL Y)FBHO*: S1"z%¸[Rj+e"= ֔bJT% 2H%MVkH$?5P:6L݀=t_Ԃ8~Tɤ8R S i^v;d~Z'BS YA_`:*tR3ǹr_EW]wo?%߽-N~G<|这V?>M`c.li:x>M+<$)rܲ]YX+#,~og"=zn|3_|:[{YUiL0Ű|3TyIs8=>(3O}_hZs{7N^]GӥqO4Y> Pk㴊\ԛ=_ >#$^tS|N+3ӾGkT|,VPviίrc.VRTkAOXnh+Ѣ(` .Ĉ|Q"tC1)շ q1˔Atw~`~_nZiu7--mыo<^*rbHIINgI`22` X&Y{e]%BY_#Ii1&d$22ӠN@].;[=A :h>eI%oxdȎZBUݲ0䔓:TM2v% *0-G!c*٪;J\27 ~RBU=\D1ʙ >i_r{sSY}dᝏ d2: k9 :$[T-%bmonU%p i+sL3ωSIm R ($D5lFB-a:iSK#]0-mQxlO`A5JIy--?{׶HrdE'p~kc X`0e׃A^%դL")QtJ-b!UQYdʼn^Hgx UgX!Hxh5Ē2~2w1`Dh[e,c< Ü@#/,*VBH'R% fcI; 7]iבpi`J fB޼jPR\Wo"s^c-BJw afs'>WN%TaҹdBic-4;ϗvNP~]{ kl1/>4f @][ qtU  7TŴѹa0`ءsFS]a Iwy.j #d`09qFbihͻŒB&DKh,wzTFN"( wB)WFFu&ix zV bʅs <@b&v!& %3apaBlo(h(cx)U,JG52⓪`ag |jOc-0 ӏ!s6,guqSZ]FʄFj  | 69 Q˃>ښꈩ4.w-c|zd}\U8/ږZڋ7Pٗ9056\ FxX@a}33-+A-t1E zABl|)#<`ɘ8w=C9mِC~Mj.uRLYY 4EL D](0p9( _c.L]z3!z cTv6\ܸX\n|Ae%3 @Ř* 釟~,LdJl 7-=7YcnvJLHۢGbzŅ!6MKӔk !o%jE<{\Zo=/d-?'d"fI_R0v|*&U̹KQHakd(ស5|g'?ݴB[ndp$7I N:iINtҤ&4I'M:iINtҤ&4I'M:iINtҤ&4I'M:iINtҤ&4I'M:iINtҤ&4I'M:iINtҤ&I)}M:iB_NrjtҰVړN-WOrPgs}AFTf4*1D10@Ƣ$~ hG^р4JpѠR-:b RLZO{AtGw3nqM/ъ=N=yKo|Xaf_#7M }Rz[H?fD$&4IM"hAD$&4IM"hAD$&4IM"hAD$&4IM"hAD$&4IM"hAD$&4IM"7,5a}E"hA[VEаRAE4O&D~:|ZA]*:q*Aa0A+^gCSo+:zVdbJ`yJR[Zs`bRX)*=Q͜=_/vK{6_^|.:\??b˱`Tj˄I q@m~'7w3P??w9<[}k(Z=*|JU:JT*Pt@U:JT*Pt@U:JT*Pt@U:JT*Pt@U:JT*Pt@U:JT*P[ڈT#DSrj*Z_|بU:xu+,>͇Ƙ?V;T6f2dkz2(+JvΪplG"K} "|+?^{.VGgG'ڬo~^[f}x~=hx9KK޳CJvbӋO竜Bەv`c`#| -?&kfD*-& $C $4&@8@aL [7e(W5aBJDž |rK㫢:]J́^/7uym˦Zɧӻ⤝nF߁#fSg"|;l.:גA~Z⽉)Ъ6iO,bb eԃ@Qof7w?Nxo xsSaHcgj {%PB@{J5Ivp˲0}T~!_Ǐl>[rG&s=dqV]~uO|Ζsb[.Nf㻴po$,vBeilZ-] ߤ7?bu1 ¹'X!;G}v6_q<ݝn'ri ́OlRφ8y<0@ޗ w -[\[h,|9p-\p<ŃN͕zߌ豟A5ݯm?_jwff(c>=PG!]+XIեF:+&*dN` fي3 LdX_&C7͍kׅgTI}K׵P _:? U\w;K69Qj-o'Qo)3-N):%RDHS"uJN):%RDHS"uJN):%RDHS"uJN):%RDHS"uJN):%RDHS"uJN):%RDHN۩R;S臟-tNr&TJ^F_);)wJl<_IŰnPG  Xdƥ! XtűiOV^+ڊ=Ҿ'l+#,z^e]/DKG)VUJL,T]A2E^W "+S*\5am\%xQ'|ꋑŲ^iesSk 9m%+/p0}഑+\ PMŬLBs&i̾_TۜxT˝_ /zq)Mg7I`VkP,1ij, ؗe*n9D-L m}*dy!9'.6m~y.bcܳx=mm1nZskmͅbK6/†8vQ=h7]_tE$aX.E6u0O E!qL¯E`R+Fs,Kí65w'lѸ4`C_1=Ļbm2<Y t#К"#^IF"Mc#u^ǩ_45YHy|dc["( ?%8L$ [H.](5WVPp0/0! bVjt):K\GI8oLh셬>k[iVE9g?pbRyϲѧĥ胟x$٣0ĻK)u|ckЮ')`ت1Rkƭ"uzY>g eSDzc{1==33Mgz~Qx7c\X$_\pyejI"ߧptq-Gyl?uR$s)d%Mtxx0;,ljfӞmTyvE_]?V +8.ivVk~-<=?xjyc1gslZv)@Sm|bNBjbcMlS5ds5mfyptH0J1ZppQ z֟m\yZ-j:M}eFx]4}FtM7BNA,Fpg2VBRhsmYXX>l| qϯO^<{Ӌ'|O0'OOf-M U<^٫OyjE$UI G(B|O롋|WΒT?켈)n-_j$bh|Ud ƨ^ޤ^f7z#g. b7 ,ot1nln^U|~R- 0 .P̄Ɯ(,i)oPF/pKVϿbC3B( y-4g 0nxǍn_}ڜwꓣW:]D_U]0ElhPW)S~z*>[ȋ{B,RFemCoX_obt]R/+#Eui=51aM9 4h! >OץH2Ezp5Ft;ʈB]{؆f%I{.zmB!+zPrg/lΌ>n0eRiuϷFC{|j>@mҷi6> zba@'0Y>1w$N xl?_slY]U7z{; PۗϏ>k3|n9h^>!U!1Y ' mОAkҠrcZ A*/>&äT7قi!j6/@94cǽE/M'Uei6+.g|6Քm0ZZ~RKlgAc왤N/6vh|`0 ױz ՛V0ucP+{B_3;Dnձ%ʸMb_!:ttUæH-T@(h> P.zX%[,6Fvt5M̷e*wY / gϲ]zR2YY>{DKӣ26+sY5mR(N"I * I 5Ak?{;^>m=Wx&r:ʈ:sҿ2q&Xu_]_moLMBhS/T $]\$Sz ooxZ^,ZEWlmbnot4;:æ*,l*gS~^U|\:sE>wnd*\&YWR0AXqY6sENZ .u5]ntɡ:r-h{V٢&T;x:rq7jǗ:;.M#|zpL+իj\NPk 0{D8̏Ps5kRD;P3\PB9oN*Zrkյ!wcFV:boc\E\/"9hAoc=XވXM*:츯xU}xA }wҊ /,Oo$$jФ@W BpSHCs\h{[-`ʒD՟>:|/[:xgeLҬ3B 7HYP$*hC<X{LD`n;Ig2ПӨ@aųSG㇑`d>DZsFAuB㽿n!!VK"G^{Yy@oӽ{);.K;֎K6;.nw\x%qw\x%qw\x%qw\x%qw\x%qw\xG0`Rܐ`;@E%qw\x%qw\x%MB=oɩߧ-`3]؂3p ypwy.|ș>2 8;e7-C-A2x` c\A#nS΄`\KCǜ>DHwaj(̘$ %hEA͘BRl9-q `Ti!!(! 4cx"F[+ |ְXևpLWUav48^3V ѸOw&"`:DLS?PhbFpHef9 3Y% ). gyVǦ3{`ƌ8K"[kku 'YD/4ٻ(}|?S=+y-Q~%俠}J%IWH8&/Qpar. @gg#|EP|vY鯕6n.s?z |w8\ ۮ~XO.~;TiW5V:x}7(7˛^Lg,w7NE/TVâFΟ5n֦MGtnE>ѭ^t,66Dmm 9۩|{{ Bw97ۥ 7sx3n/__p>"O>7󠐛es,zk _T| RdS~jL;wϞJ*tbvW)ũlHAL~x[ǻ1 71"mipy7|y0hrAK͙P>j`rE 4Zk`Fħvlw1<#T $ BtpG5&Wx˘Qm2cFiw!?acGHE|±0 "\b͘:E )UBkC0ԂJSr-,-™ ێ)֜.\,X!S$c"X-%KK 2 a($Z4"t.K4g::{X4(p0a9 F*"%!1VsKXQEwwyrpVR4Ek䡝F8,|sS;"*]|^ YD)ne9H)ȕwւ=w#(@Gz,0 8 ItQD #)g%vqt=RVEA90HPALav'[QpK8 N$+8q#eZsZi^ d,LAGX >HRrښS\KIYmOY򴌞f|f?(d?TwY> )t?p4>ףlqjvaD?f8̍qCҖ<ɖJR{ؒl6G]lt,57j¯ 8w@` ZU#(`+R] Bre`٨K2#|YO3V_v>脢iJM?OT0ʻ2g#hhF0?>A~RQǗ_ӿz@bR<su%@2p%J/ Y񖙺 XNTl '|-YO/6(#`uCoxdTlMVI%[-О-fCe>h_͋__EG~}˛ln|fM !P]!WUWc1M5XGP0ˏf?Sb6L sX0b!uF [\1/G|Pk>FZrTeTUáKCOIN)g JI.7Idӂw 4xQH,L9 $P 0AApnWĀTq2+ $`M#Żp0Dp(C\(wk`*䨰i.Xm#w!Fkm$I"6PD^€lcѰ1Eddm6IQjJ#;)k0=f2NSqC &_T*QET]ց^sX?eXyy*iC,"edcjʲTSsloPdcic#cF?Ӂۺir$#1jm;Zٴgv~zOe6~Ri'}ZY R(MT 1{ͫZ>j*:RLSѴ>EӶLrZdUU0Xh7=NL던 1#W-|ll[n1,vӁe-CFPD܆mH9ybY?g13Ec_̌IYM]8f't*A D#W ZyTd|A}*fz;!{4o {ԫ^=94|BPbSTqUT ԢU NhQ%FQoMдڋTj)2,I$D΁`HgC6ZήĻZZC*MyrP*CV4kƽe ~%^^툥Fu(1bA蓩Ƭ%;djӨKM5&Z)} 5的)9E jfх\[k %{\Rp#c? ͌C',|WX2scƧOv3-.x9[l; pb#ЯdI`dk4*FyQe!*쒂/V۔me PE۩@_T!vFnُr7$XP jYɢutUIP*YVLXmtl&"s@i-F6Q 5f[`Mn)EBKKRruYm DpJԷ(\0͏T_D5!℈;[B"Y_,2ښ  eF]a3>LI7)e$)QĠ ڸ86I9q:RDfQ[[Έm8Ϩt^Q']16KE.!FiF{tx6BR)L ]`M5M[;@p8<6;Ƃm<ul.?Tˈ]ApSE?#lꁴ M)rnz:fܠ g,)'x;RAcꨫR&Z xO:hC](bJ&!]I}Nic#bTΥ t +h_zFNc\Ec:HoS΂Qd=8 OQ=Є07bS`~7i}g{1Q7@sYb֕A>~ݎ_@?]MQ m*5M'aSI,э(b83G}&:f r*tL{pJ:\ЂUVhYyV*L? 癙6׉x1x8f=_buj8Oݡ=9yy~Fdͯ|bM6kf+:VPl,Ep6;H$RS@WUaUB_PkHS$T#-C0ƱA\Y6$Yo{5xZS:1IY!tW+(=X,1W)vL! .j=5APʪ$!#!A\B5kP`s%osMɓVUDhC lZ ;iIV+`42YǖG(,N rP>kX SH5O@^(*Ym^O@׸< T/?]Íyٲyp̗ n@^¯g hliCrGb&F|5Z寛}l x0&Y"iЅz ZY 6=u#'0:"FDd8l]02PbeB+Xchu])B:rb bUKމ_'^rQo~tzt6RƪdK!R :{%z+DQg!ToֲP!N%2.:o>5N Ȉ$vP2{u~#g]%cfu17_V;iP<}0%/޵Hԇe%Q[&F] Xe¿}yIf8+C̃5**BR=RjPrT<4G"]K%}yn.GS]u&jlL]ϥ-y"vTGU=q֝}rvd^}<;|ZVCSՔF9 L-Zpƃ˃-=_ͮI,S?orv_7.tqq1]$j&Wz$Ͼ /zX},.13Ax%?o\ǗUz$x]_ƶ1F?D6,ιU j]^S+>ucͧ ˷_s2l'NM;y8vλ_k.>ϖN,!T ~^ݸ ZPmr?| VJeߝntcچ?o3 k3&eǎha{Dݱ aG^hxq\M ܶj;GUnrcBΏ<\vh-6=]v/>ǻһ sz#iDݖkP`O~4]kX'ņtn`zr軐+6-6HO/A\e0&[tA+34jz1Ca9#3'g]/%׼0GSֱ(U+^bxQQ{R)Z %PٚVm&\k3ٰ(Wcmn{3Gjӑq%No{lW՚ >XpTdі0mU}0Lʠjj{*uuueDv,5q^:$x2ZOAQ9K8C%d撂T . Q%Mez90u4b=}r2lORk\V|_M^">_~o!0mO=gcU\e}=WPAmaRO,Jg\J'JFʵp4Ju 8Ԁ,slكK H1 D'WWfRåQK`~<ᱻ E[ͥVJhQVYmECbMZWTLEy+V]_>p_^]n}Jf+3?Qo̚,^?Tq${Ďވ.]\-Y 4R]FI98F?m3D J<8Q@@2%<)ɒ#XЮ%de kdp7pӶlqkE*YURb3UcX1eknOИ"uv^:}{G`A[MoĂ^'Fٻ޶&Wya XNf}ڂeɣ#g0})RE[XqM6ͪꪧY(ݍWOM 1B"ihn)1H-h[# =n\bREL BІ M7$6r1DĀ )uHHJnW^8JZO!`AZUy0W΋RrE/mȌ_ZMGXOО9c1s uLE"8rOpːۈ喣)5[M\u7 jVؕooԖ$t_ ~gg)%NsMB2 U6\4$&  aJzU%6/y@*=*x)D8UF%[a,kYq ןUWV&Uw6<3Cٛf:e:2J"]m6x7N0p:B˒%1JE_;}?}y,QlCλ) e03$XE&FxZNv/(p \@n] ѓm"ބ;t2Xr4X\m(?=/^W`bmFp-Y5QSPߜ%I>xy{qzܿܿ8zk&.?x&yHSQhy,a~m)̩^c>;Ry]f^WGѯ|#y4K~32$;_aCtcR畑XA! ԳmGen%Sv Z;}sܷ;E"[R0YsA}sܷ;}sܷ;}s%8֍7l)=ԟ&3@bh;ҷG49@nA77 دlwmv[u'`ipj7 RQ fQ\`fІXA t~Yʙ$2i Z!=G<'bE<1,r F؋FH\qZA3paD4ӶjN3 r TjQiBG 9ޑs ĹYݮҰROLx,SL L,)X #aV<̷Gi5S<<" 8'ȩbLY*M%.0G6 RepF1͓uO [tCqtYs|gk[,]Ȳ6.e&jFlUJNA6"p1IRL+uմ/xva EԽ=0l^jB'.|Cw&O\?UW-f+0no0wq(?Y,t4z)E°Q)1} rŭߟp),gzp,ior67X|ZRpFWO R9^M{(^;8G3u8_h8\T14~hP]n0iRp4UNAp*qv v&ayuu8.HiΓbp׮2MvȎ3A@~1@\nq1{= )Zn Pg_o"8[_]7\ɦ͑=)9mV[],x?p,_"CO2+kί]Ķ8{V>=g%y ZP=<ϬX[GѝV_FRlf7Mf .Vu wzUN^\U(f-gTjS 9E@:0d5׾~xR,j| ~7Tɧڹtpepݹ2euۡ`|;Wo\zJnslgܠ -T爋w 턷|cR`q[:e+;ȗ&m}װr6޾,!܇wAIJ73abZ!4;A<:"^4`/$F{Cl>odt|X%ؓJl(硁o@^C^4`:^f9wy3{͎pDrXbBat$A "YE )ĈIьKi>"X1X-#7K/. ~{a څ.x yh݊ht_=V|­ka/wqx C\T1͙2^/(6 ĝ4\˙P.7ڛ\Ȍg(Q/ !Wr;bG цpzBT Nv*"6R* Pc>J SĘ>X)b4(1˄bm^N=g]e:[ntnTAF( ʹ@-b35QmE8AsVr㉝Dwܤ8߽`_eX܅UguN#9]*cc|Ϝ`{nq}<#re A%<@rFϵ,wJÃ&s!n-Xmҽfr= \c7l^Y֘^iGjd~sVр1[,} *[kݷg5MF^3 pB!Ydh+P>F'VI.<}ٷ}p>^ j D5>[:MVܳIze],UxV3Ԓe6Pw4/-+ ͷ.IH&ᚐķ~!2r &l)*bb,(3xn2肤T{G%uSBjG)'_G&(Cua5r& D `kƒ_ j`P k8T,w;db|B^\-ەVTxGDJNcRYu 3âҌ;$$" (O a}ScPa(~0 ۰ )c?"c 2"I A#zZ'(_@mIPt*vB)fQ-Vqdj ҇fi .%XsqNN:FW2 JpNڜE8ߺg Y߆lh# ܪPDfpf/1$r&*mƎB ${؏5<5 =?NPCuD?RLjDMhkYV "aHuNS1qxJ:/U+X ΊN[54rTa#|cA 鹓2Y`҄||v>3 \y.MHY}Y<>N..Wbyb,.14mq[.WCK ց7mZqKH11[S:5OiTōWYyMj^I P\ϽANfsKbn$jb'͙'/Bu=qY{wPm7!$QFSZɇu5E6od]uce #|8C%#| wQp6YwqgY{Y_*-/^ngwy8yNo?zu?U_0RJݸBF͕^12J?f?M?KUmNQ:wV$GlCY1)WNx$@!ntѵlk*E59r·69~[7Y@~݅/Tb7C!T,"wsgW !q?I,0.PQHƼh`)3Jz M6wn/pK^H_t8j1ƍVJ*#e0"d`X5 DA=5|̷sPb_-W_:6yتYզ "cpjrQ\G),ºSH؈ 9|=MtPPjnM`U)ӠM~u1p@|>WMCz9jﱩm":]I9Ub>%'^m*d_]jهw^hmn֋͠)p2@@D RYbw _>G?u`[ "ҫ&l_}^]~X^}UZU]]BՊHNAXgp,N_Oe/?nxQɭkC"te)Nz'k@" -G# Eٻ&%Q>שtsfN$oÇnNZߌ5ck(OO"#r>wWêʅ_sG] Kۋ/W< \Û 3iU h"=JWzj0 5ykmX_] X:;)2v2'uۓz$ -Y ll۲lՠ)ѧ! J?~zsUΑV/fͬ)ӽZ}O ?q|Vvu]^#;}ah"Ҝ<(^g儋<*u&{[A#"iM0GEV:*20Y{-cLAj$;6@S0 [!)RƇe_w@[39D T N)Ft}Pez򍦳WZ'b 6)& E"ThVFp&uBbTIģ*bZX%V@YJƅ.RYJ6S<Ƙpt Z@l}D$ЏtIִ=ߪMs| APjI.\8Y[Ȟ~A2nL4#OPm&4Jі24~jJ=_pWm3B,[-ؘ33 E:HM9*Fib2t/X&3:gU)ksI3' ycZs.Fu]9-Qq`ɛ,}իl;xS@6YR 9l\1A=1#~i:`<19 B񌫚V=| G~EH!8+BpW&fW,fWAA3"+tL%Hj+ݞ`0l#3&(!rm`y.dtGp G8cYs9#pmksWsvtI¯ 熱y =HpCZ.,+5`XPRߌڍ\ip'(ToҼ@y ǣC|>M=HU!l:z/F3xwj ο"`D9<~PbʾaiF>F'̟ +x_~<$eVM$O*Əe~_ mA}֤ċ]5YK.a Mlg5<[[2^MGu++>WzS5W&"Ɋxk_iepf\տr>с^+ɚj|Rs@i[=Y.׻,o~jײLw_yg|4UO:Le9:8w9r\{gk[HO퓥i]:Oɀ1x"$#99̢(c `0+ᓳk¬kͲt?|'x!hp\1X^Q5q"h6;ZACw[ף=}oGoHY\>-uQS/_uV\\1|ٜF6򹊤=vy?"I\SOm:iLhv"K1p̨o)^Б<# &LיC wNdJ<@ 107N3ç!?OI-K 0Hb<3Dg2sNxBڐȾ7qvh4_p +_?u{ :8H?!Gг0}/M&N8ZΪhA2 NǙ4i@v1lVP^a`V"m2n[Rh'Q XǞq5o:"~)9m2NE'r29SI1P9;íڊ&4w6hwȠ6qRRڀ@#y.v&`H`D1o@w&a_lu s:$_god_{pGiHN۽S)ȵ71L$\HB#O"'Ң3sCC|?{ &l_ӑv |௹V]cя/X 5{+nzsR+5PvAf9T;1X Iԡ9)Ka*A @-W^*B  7sN(o-C8;9~z>V"u^o!]5H?VAj[`aGKUvk93ƪw3(Yyu 2i mm:%iӶi3 ESO^_xL?EhX3༬1 ICy?d$).~ &W2oxG~ ZSMC` !|ʷ.fhv|nRI/8OgY'-WoP ,o$6]>ٿ'% ˄{E N#(w=Ylnlkӯ9 Z$ Tp}%QR{iv?~'FњRmZ|5.~n0΢p5itz6zt;G͖Ic6 nȚF.O>nɋ3ܼJW7ͻ¼|-ϋ%ߕF_} vҘ$~"sR/O[_49P=xki\*EoJmp3=ʜ1~={F9SSngi_ljY 2\3+ KkmP(7:9RҘ( mCre sBqR zPIcþ{g5=,]٫J" ִ])KЮ3˻_K˩wڕ61c68&0 QK{S P%ӫ`-DgIJu6IHxQilűHlC}{S}u)/xTCRpeSY(79iMdK2a3@YVh6:9nwZH4Iid44Q③_ZF"-נ3 :gG+<5Q]'C 6mlH 6\[NPq3G8hW˟1w O;Ag^ j::zs<t]  .Cf{ ǂ76ii$O0ZȔt΢<$v,;XdFtI8aeLZJ@Zq$ېC\؂= Mb9tV1>!keTRlWLWjg0jVj㴂o.:zM*Mc2QWuT_~4ZUSpj&Y7URj)]UP^қ6lo|\o޼{S$yzsx/bMv5*owቧ^vM

䣿n~Mnӻ-,4W>[@U߼흺!ew<@jaI/Jl[wp8oK.UkMj*Y:LʏjЉ`M5DC|H',rOvYg:&i(mNO"ƣ~cVW>r=2_Jyq~6p6L8Qn;[)vF+c`3dǝ=߰Ri+i 2nv2e28b'ՉotO)Q˓4t_bŤ;S<1"sG[/b)8~e6?a?I 1:9; ~!}$,ٺgW k]w`in/"^ Kh x!! Z&2ppLvC:A;lppM ۂ$ qOy%!1Hd )ZJ6-Jo ۢѓXV1fL 4NH3p iMފ 2K}&#Oh2`: [^}oe9sԎуjg^Kɋ./>_I R  xI,I l5hC(b΅-JN6 )bq>Ip<'!@Ʉ;H vܳ&v%^9=4\cSN=^n:G7{U=n3sYf;#w}!9UrlƇau ,R0Z^3?Y[Ȟ>xN2nԸ৲2 OfZ0FƱhsЖEunJGyS{ S,͜x t;p-F+FOTd7 6dywoiv]M7s!4l$uw?=b1;EnM// ]Tی !F)Ü%\x+@ř'IQsiЙ%P~сdh<$ɩw9$FɃ1IȀ92w혾ٮkq@K ܕ泼jcazN1k+򅘥zg)&s~L'gg~i{gzF_!I.B@pd2'd&38Ahv~)RdQeʖm{aWWյ͹8gHr F؁" x=!v@!B0B0G,f(i*SHh`A-H+]QLm#aRNq@qНʚ%{gqoAUmlr29َa \sfѧJY~ 5eYcۀ,txj2/PHo8I2M%r/x1^rÜ9QJv6ƒD_ekgNtk([utg|,5Y_0FQI!PZzM4ԥܩ`U@FyaʅPkƉcؙ-jCoh<>x{AF*Kij^v瞠\ArjyoFd䌂^'WCsKFD§@CAGA 0EL BІ,F|@Xk#W_D ` OCBRuGu^9s\hΡq⮁̘| " OО9c1s uLE"8a*f텕[Ŵ#VBo9ͬfmrD][;YIA\ފ<6x1Po-U,J0ˌ3˅6HԹŅJ*tfկmɸYR?|΀2̕@@w)Rs$sB-V# "7pqpWAuc w4FD1:N6)4 '0A;I(&0V)-G=]CZa=$oqp7iǀǭZ;Tݼ[ƣ4]꩐1YP= t"-$I%(թ{޵H0&NaG&)$ UB`2"BZ%g WT)EVIߜ'at+Ly==\^7> t {2,0F_da?b2"4c0_`:\T*$u٠ HK.9H twRRz\/sˁw\-y6VU5Ά}M_SOuuPdi.|t֩O ྐྵ~+ޕ=K'IVq׍qr]W z( $[x_i(a=vZ㰷b4pɹ gu/ z_d̜"|kSڶ8{V94{ogBY`|fMn8v=4|8~ , շZ'Pzlꋵ_E*ƁV4 SycȠ ;z 랲-'ԯj]iD}j4ռMn4dұ E͕UcY3YSi-O_^txջc'm75`0| #8?6oRa߀\$6DV[Ub]k͊5\n`cV4ជ^n{E{ T~{Côڛ6~;⭞xMԸ|DRۆj|čN> kFgZjN߱$Ac0(Fu#t")yt ETzmi^HҊ||X^B ?>/Z8g9zE2QDrXbr@agIU APﬢFi|XlbĤ;1lY>ҢyרԼX2o-agWb.va-ƒU8u¤ړs%j q-K#A^΄rDfoNeNPwtݓExo)w}Ϝ=7'>tBι`"E?|2HUYGH(򹖜N)bxR6!RlJ7S4KA.IZۻ,1}g[jlI J#68R$#Y )Oc.*5X@Uo{٥ub ?å|}m}<%ؿ%8ʲ_"|~*^U|*_ƾpeݝ`d$?Ȫ.ȪKe{72k{.ѥܞEi R1L 9WT^8X̵Q*g er+I1TW~`I](أiSGG&(Cua5r& {#vqy8[QY8R_g-rͿwk8نpu"8w^TZء{BG\>:,xkKXd=e9C>Ol X!XDZ4AϬ) X]ǖm_EYs}ӭ];7bS:GnO R( LqMxpQ HMXSN˯9^ yϾ"kVoyHFǤ"u 3âҌ;$(" (O A>_K1cR!Pc?R@l~DDdD4 Rg+ES3kӞxm{)DJ4mCHئN8 -9D5@H҈#뵷VS>$-H%]< O9ﱆsq1)̩4P*WFc{pηYBVa!9w񯹥AZu#"f8^L DӅT,ga QadbeIJRy.'0RWM Jzxuãf+3ў#@HFD0!::RLݑÅ;O ^Ϋ{ aDtHva`Azce,bR&t:_J -T+h<99\ lyX\biqj]_Z`G tF)Ȋs!aɈ0y-ͩj͚'|-T+oNΫ oٚ|0klm\ l=k'" 0*PQHƼh@%=o%M!NOkC\z7Z)|ӏQˆ4E3c4(xo A\*r!{sPb퉇-΃[OzEe,e$Z1d f1grH8+GXzacN! @k#9oB!gh՞4#,RFt, eY{d^2 l';D}}=Jq LV-ۢsz-B=z*\ҞNz$_3E?(-,Ø p53۴lhVG*x.Esߝ`o@K ]LȚ! S)-z>E6[-"ջ)yS[hR)!_-q垯pV n6# d:*Ƴdo~*2{e]mWMG;lyl8|3ˊ N/(_~z_%lQY!޹h H`n!B'l`\qӍ0r&uHHPngGf5Fw#/ 燓p<,,D!#DyeEgZ1j974KGHcOF±[\!\a͘zd­Rii !',u $qK !!=PrIo-Hkù)t.͇I,X#ISwcf22A #Q@:kRĎXA,0s^h Z&h_" u-ڤ1rv_栂e`bzNmCs:x$:CYcyoS?KtkeeJ?{WF"f0)Á`w1E$%ϫ!R!SEe`[4Y|]ջKCP` >J"" j=ݠڠ`s ^±X,Es6C-l"d4R# ::7N2T3P+ mCNUFhĩ -XStQ,.EYgwTl!Bs.x)ђmҙ5\bB_e\& |1*??h˟#fzq=kgMZjW.[򱒕V#sT9HpWdX"Wx{8DUA3`ڕ7|ea9FrLe1Kеãp+U Wm= FN/ D҆fnex!]BiCMT`O!_.y+w)wM9T&pJHmL` OJVQ3KkrS:[h 5=8Tb*>GUW߻#ȃ)qIQ.RK!$\|4ƌ+^ǞR:}gQ@܂VIH#y yN8CAK!D ;MpQyHKio*xs">ӊ+^Wߜ-~hni_)~'/Wb#gBZh섮٢_ħ*$)IM"HlArkH OkzuX#[I>Y,}-EbPt:R=;c`13O8&p*HF`2gJ˔QoҘh:gr.}<pKm圂d"CpL8Д& 1;iK8k;s^pQ]tp~2&OJQƗh?ߒ'NKgULRAdIUӨbV`VJETÅB:'*cІn[KÊOڞK箤T'Glǧ' I3- Fe&aXTfƱE•Dfoi~?c~E7~i݇("8%(˩dlҚ9IZeRLe],+=/B%M2h wr;jptblsb(Q̮Xj+VVG{? 'u:ELU:k *$MFIcnPb2"C&EG ښT: RQ#p%@%Qo=&a{ؒE7\ cW,bqEEE_ ^,\tٙ"93*(>PyyW7SY$HY.t H΄E853"Isi7LҽE&*]d|,K{fɱvW]_Q eq\YG DEVԅ.D?,84!v]fGWEO`V%5S#ZyF?^۴B#Z9L.> $eSd4qa:6Jt=o㑲sm+-e4H0^y RN?R|dU R$)}ׇ9`% *DT,9f]I?VA 5۱"XrwOy;[k~݂2@my|5ET$.+'yjyeq__ej"& "q%yIɉ:F͆bu[3am !Pɥ/.lި+e(VփyDZH2Y1XJ Q`coFdWM[+2A=c!m|l.6_k֜s+h3UV*܁MzZ"sJnd=pz NȽMǨmar ?tԦC1oضdz ^9?AڵSd55LL;m ﴥZ@JMs{k$'o!&A&HWx5˔2RmݶtMw-;vdƀRnAxqҁٜ)h+8S&'pVZj]<钋a*J14G|N s  ZJ9,0gC0mJݵ9F M6,trW∊PASR\^ڏz-z?i=h%!3AL->^ǫ_[=3TV3%XM$8m$1F A[U?{go?; ?O ]7:|MYB\qV'Tgr@V\ R( dlHكH 50tsDu{2&kLj2K>q> nB|yŒE mwhs8&gsֶ>O\>8ŝt @ȅR8HU9i.hf*M)@Ύ1g> mv*z؈7{f *j2$vB\5qkV qc/Vz>vN1V=|&fig)Fs~0M>(t(Q1"$0 BqUBBJq :Wv,*ku6RP@BsgR!h mXu= l؅(!Kr]d.T.$Jk4'js.rjY`vyk׳~x?qL%m2aTSXT0ua&,ALW ZCUYj葶\y.-AΩeoN40*RXGEdOdUǁU@("$(Ie慳VI8N6%Ù̀K*鐠|E\}\C&8ߑr3XmKrqӹCd4%.R鸠d岶*)-cyiT޹}H:$M r;PsԑL`J{|OWMJB]]2XÔ0UxDut#dtCe/Zuk*PMRԇ񖞻>wQ⮋tBZњjd+xڠN-s%˷Z-t3_/g[g٨Y$/6Z=d4]fAs3B_<ݻk} 6{)8±(eA3緜KܴpN~Ƌ`py-AIH1! x1DDD|.*UTG;m{S:|mO9Qu`OeڵVe^̥k{ 0 ou7jlBj d)q]<َ"C,pCrRRDަx%yӹ3\Ph{7>kEfB>}wwʔW]DA`=sb NEtqza!k1o'[io?/M/m2r`UsA]{oTd/gߺ8&8z(ͽw29uyn4!9$Xg홟ڹ[Kia M<52cYoE)ā! (&d [KwG%eFE0BJ$ڗ o*5} j ]B?<'ûc+Ё S--+0HZG@&En~n#$\^Cg {pwA30aD4ӱjN3 r TjQiBw\u qA'ɻ n׏Sw{+xͫ\eZYϞhGCt;!:TJ>9Qh*}sPzayY0Hxtuς.1{S Zƥ^FlIm- H[td0T{o!֋E nt8$ehP[)E][v}:@v/uW.w/I*:x9edP^"W2"\DH x*l"+$yioƯޭ}en\u _oߧ)RòzTԏf8>d>gs*"gO'ߡQǦM1=L/u9+x}(xR(A~`/>x4;[xT\{hP?bH=%$EHJg5\Gsr`>fTKM·.FU3>ڦOt4Ә[0֩1s7u" zC44;pa%?^?cbևkpZg0!{bȇ޿{W۲s׃M{ZaZ+؛­6}>UpV׸i[RENҳfͳS)-Ww'l02>f` Z%.>fǣ$, *fwyWԈ ]MK\H\ h&0b A3&ϟsSj>IJMUΆI)SL[Xy͆~e]vZ-  2Ř14ɪNa<}v"rtQ[$M5ht g;Y v 39fr߫Aa֋-\ ˽/&Rwh6bۛig s?bVѻP3h7#ު$&%IE2jbkֵ`' f5LyNR1 k H:A<:"&4`/$F{C҉|[QB xQBup0ݮZB$Z[U@(`zDP% RX *jdGO&FL'0.;U3ܸVn^fצ; wWl[m%C)]Oߎqe緣k(g.~쨞3:9 [}J71Y>z]gyMX~z̘hưuʈ(FQ U+]9/6<2@qfB:UR} DJ;f2Rʃp-AY/5^#:VL;g| 7*Y$H.+_iqׯkՖ!n>ZumvlŤc3:M2A=@A T\)RMMXåv,w54_ux'Kb>Q^%")9kD*"| 3â`h%1@!Ay y-pcPa ƾq@ ɖ(cH:,dHA L1WάN{SJ4mBHئN8 -9D50jGko@}HZ8{Bꖐ3<݆:Rp"xS4:FO2#FNZ#o[B!F_AYYgv@$uKB81pq|?))5j<&$CN/7%6nd|ɆK:γ3S0j`e՟+LO.bq)WFm _ZB|ۨ8wDH/8&?fTɽ 77ŋIrZ5aOʵz @᯳iQB6;g JOʑ_=i0}ZYD!UGIz'1UӸuTdۨmse@#u\40e4.g_1K٨/STNYv3?;t>~0QG/_<9H LKY%笴4az(:|_'.۞4;:y ̕~>W`\Юuač͗2 pi迿ueWCxT ƨYW ɸ)Iju`! t)' zvIt|P)0Rz6d+ DQcP䠰֦@!t|EÎ1nRRy쬍Qˆ4E3c4(7 頇{.8E{SFĵvpFM[ xNs4BڇOUZ6` wf gN(?̎\qXA$;1d f1grH8+WMLj9+K%@oUֶ=6UzFwhk6!ٗ0̦./ |(T^]6,rˢm˽si82li wwۦ2]Z^Kjϸ. vC a^fz_xM0x36(!tUlE6{UKUoZ>B՛MdF }m6ܙa?edaRƼ9&ݤ)D\zb+$炉 `^ D}^"1Tk⶘K]1(Fy0̪7Ylv*NRU!$8m"BGvLY)u7 ĬU۬[t6NJud/{[ꙕzYAewsࠗFq(Ӏ*ys%AS gG7e~9MK\n`\Kr1otV!=7DH^h H?uh P׭i#oL!j6[2UdPj<sJc='*{qQsyT?*w۫k 2)-$ VFch#|NմGgtb ,%<?b><V:^'ޏqf2tTs3(1 i$Z_98M泿ȽGe⣧U<(A x{l0E6k]5޶d("7w?dbCn/zS:7NXZ*<;4cDIw\򽚡/2P䷯5Y;5"n2E-69/-`мjI8;ɍ LGL)q4!G̙qyd%*nWKh_ @B_7MZ kIør6 XԪ8ՊaUnEZ#`EʫƲn"0a2{+=/͏))l (,K"G }׿`u0pzñ' lN(aCRJsoyU;F?x}KЖ3q'&E3UC2#H4!q^7j)RQ)\İZ;_y+52E^`:`(E#m.8OIc%¸9Mj[lI1ZW8`%L+bKhe)!D!e!x%jИ Lk1hFs+%ZXZEl[DvTG byK#dDd̩M4T@kI@- n|ցbm :fNhXj[f9Rg7a Fc@f~ L{ M%a뱄 `uUaJ}{t2"{pè3YeIG/oYjSܶ-d KB[\C;a " z.Fk9QJ*VQFօ5Enuľ7:%asqʡg;[Gg_)beߠD%LȨ͠4vWxAEbTt: ]EƦ !r,2f%x fU.EGV+= JAuTv|Q]Zu;#tzY.MV !S.iE)1٣܁t@>1(`-iallGAuJo(C _dp/C^86k#eX. qah{n6EgFxVjr=teS`!\Yc(Jl@5lmxO1ԀMTmz7 Q we F8dj3ER).i8!o*2uV̙&t6 6pHVG)i0(WZBJiT PfPY~PHXPLf6@&,=7 !lFCmTzJq2M!l'@n@ e %"tn|5:Rb" 謱Y䑰ha;*p\T]5scy\ ռ @Vd@&+tpoRMp%JE5nXBtS;KQ@hZGBv Jq*ͦ.o-((H}k G?FASɊ5!18 j]g!BYX*k qq1y{T(&cDXz)y B78m֞b_2 Q,03FB.*C; B =!^g7T4]E3PxUͮD@JD 6f03i<)Ԅ8o _nQcA "z[!@iRA2QӚC5QKŸ0\.bA^!HW &M&[[yQx+ @e6"YԏoT? _Ey՝p6jue:84x&o~ӶإK+1ӠVLA+p-F$2Fzt#Ax( 7DR.sI9Œb5|3 hcE/5G"0Z]l%;6@c0 P(U_c+@s:m)E1'+`VQ `:('^l%8o,,$f9j&@?xPZC"N7o8\oF_c,I°c;gũPB@6tf]@kV}¿?Hc5.:ΘTkth0J 6H57筷:^`юYc-C[v@¦lְIw |>QMVsp )7ϥ F{ e؈y!znszoO-MmZ~H?˹ y>,2@T@7?4F'eknc0G0;kţbq׾cXk@Id hFfc/`ʏ? 0#2k8nJxKT䰾Q. |FFsA*W{T@=ƁB0\P=A>8 =5zea8T>ǟ,؊bPDvkʹ a9w(.X Ӫ;0-FZWQLFru#zIg5(d )clchap `N}7ƖQTEy;kѪTVm gң&g&s` VWՆTBB`u AцI?샩oyZi],F Y{0m@𼱊;kE5LFZvZx`ԋA zȕdx_HmO 0%7aAFг,`l (0qK>RZ7̤3H`d-7 ѧâ>KJUeril" jEEYlV| XO]Xv颀k-0 U&Vڙ\78];R%ۣ^|oqp~]l9o֓*O9;0Ƙ'y~wrmW&"W&f1jʼ[hs]17g76/Ж|g2ܱe^ t'#OT+O>nV@J JoVY@ J Q=(|j!"%T=8% (D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%+T9ϬȩQ\_^xsUʠE @އE $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@zJ x{QJ/G sy1J XkݳWJD @_O}D8G>ϻUiU+7>k `NrƎ7{?e>\`紲|G{u=5BZM?ݷ߮e|o:Jֿ*>Ûrtpd ߩ|m/$MilSiN(Ttt)4Mi"MiHD&44Mi"MiHD&44Mi"MiHD&44Mi"MiHD&44Mi"MiHD&44Mi"MiHDi_4-ˑ\^4 ֺi28ƇTe`'J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%Q]>Zh}4yˊ]o/ɹ7xu"LNY~HͩKՃ`> wtr&tJ87*Rwc7 WϟRw{}?![fDUA;z Ycֻ8dJmϞVz]ݮnu6O|SٜѸU|oxoLDMAfJ)TzpCړN$a/?g?zzqV ͽ]x%o뛥ܮCCѸ@l\Ӄ~_x㈟;?o {zȠv&Gbڑ Qb*փ)wQyMO?ܘ wl!DƥNnwD-zrܤdAU 3)>^$?.$ \%([,Zt"Gqb|&|˱Ml&,SW@L($ZB $Z̜!YA"oDtOkxǣ!#k0kTPEPAV |1);W:}: ,T4SgɺӮ Q7T7VPz=Dbi"2|+C@ڪBVEAAs3BN<<Կէ>ޘjori gd_2k-&ƅJ\tå K7BSh|$x[nf^s_!&X(+'77ASQKs=ZNȋ~l%\g3?~W?< - z*,xo{GoW4~":ЄQ] w !JeZnB=Z,Qɚܢx x1snF/:,T7E;ʧWG'*{|ijsz"aWMy/er%n{ZG#‰"kYp6¦S7%40inxf $#/oX$'H ف„AMy&x[2uJNwⶌz&VIk چ<,-^{N ? ߙYs<@uP?xǗj܁O\z܌Exw~, ē2mv=9~Iڒjb@ 7́IMLGhuaDn1z$|o]Wgg칵l[c}U[(#dD6yy< ש߿& IAc3NjV*\z˙R hU3-]w{T_,%gKh1KC7u'P4+7ȅIx˗Gyڨ(E5h!:|LCR%򉕭>^j%=ov,ᇽfǛy5ٻ޶eUNnKy_$/ hrP>m!r"Β"%ˢ$t,'2X&Wof\^a8_ Y ?N?`#XX<<]+8'wbNa8y3!)zφOШSqڴ)ӼVqwR'0XBF078].47zlP_Mg!D$HLY5]sY+Eٔ75Ffq`nggWGì w9XηcpnS\I{+AEB?xe:lk(j9zP1do^"$[xo^6\dȞ8mV[Nx?$p`.CWR"tyڅIef͞w˒33B`d|f ]Z9mpq:Ųn Ax\l5.q$ЙAZ C UUUj{T9#)-"i^tUV{1ov\۱XP(ƌT4D ;.qэې϶->o5MLqf4/= ?&tl&:~&i&ӴY3ws,wh6,6y=iI6Xc>0InmdZmfچΠNxg=աDeɝ5l;N>E45Z*S~yal}$1 k H:A<:".4`/$F{C!I/Ьbh@TYimצ ֤a"zxZ-rV;FL;(ā.EhVxFxm޾vv+'b oP6ݡ]j|nY͊ltW=V­rZ&};tȎ~h9'gGљ׿\C6\҈;i3\n7$P 'K?\*鮸훐j*W76ߝcrKnC6%ﳫ0̦.-W}(^6Lrˋ|۫?ah_\χl{ 9keyN58^j C-XfՋPsaQsRz0>ItZ)ֈN3>RDc'~S_N?EJƊunւ^ei3:U+瞲!kϓ$Rz?>Al_;U1BǼ`':eD*R"R.o1lcS, HXGJlB &Qc(,<gkYa2h/.(joC$3tks5<{˭jǀ-F=@A %ׄ j`Q kԎpXA`)x.uҲ]kב HitL*"cMfEwZI%HD0 AP$@QcPga*#oO샰n1$H_D$Fa!C* iPTР`Ppfu(rNmIPtF*vB)fQ &LZiđ[)HV u+Hlb Y)8`b<«Sai3U^#'H%o^B㴃lc{{iV"hV^ 1INr4]$&ɞY,%4FM')PduۡN ~)&7fra!6#1zaH0 ,oC*0:jTUH+h8'ٿO=ƭ~!Fm{V2›QjIs`Fe+f)~{8U[ zT5 VG޼}}/V#ipi\F5J|2JO_.? kNOsuJ?I>W `/Xuܬ։߶(A|] CSŶ.z6یK|qX7Y@͸—*D !QM,քY&fp6V|~k'"X` \$UHƼh`)3Jz 澕 68Dw>wV.y)=ǠÞ7Z)|E(aDɢHÙk7 i{.R׻Wpxwyn 7ͪQ䃁A9Ae0W\qX*E3Au1bҹVpb#Vu@hVzN) MkFrp)BSϘ8P=d:FY^10X<a~Oy+}X7פpgC am!3ol[SL=FT =J{ᶱLǿ2mά5E yvgkىUeu)b@l-"R%w {8̛T:/Р 4P0T+6UMEB;Nx q|3r8&hV1',H٘1( +R1t2\J%E x_|,T6-ȎmxlMq%,qHBu\~LX :Ez# ^l4|fuPbi@\҃Qu+eg`K9q>|J(0FWuBQ[Cꕽ 6ۿ2`N9)e&`"Gl/Y)q^+E FMDz-`>6mq=*Z t ¡a:%>ᄯġտ H hPREqhHIڽhG%ړ\YC(߂ >{hǟ]V.+/%$v_8f F9er}\>I#88V{ӓ9 | hħ-IU So1J[**z \ o8tp䄾CRx>rs&m[t|ȑ(h$E DH2T(WFB)xTEieoa g`!+YƫTSG1 ,AFoGi\a6/GdTaK%K"2W.V25&rwT,WȏD!ܚȨdcbs#$ZODZ,r. AH&i&-k|V[tH|D.9PtZTtH`RGSJmΌ9,j1OaLpb_Nz`P1vK%Jݟ䉕(]\l#YZ{nWrU #+^O_XY@vy,(s+sG#S["?XYEya E莴f :]QlD$889SI5EB 61{or4ML*7H6fRu6B*Caj3jX$"﵌FMFS֟k:ޒꌜ-zkU>M'WWrj@>qtyhӒK~G0j`KS!|TXDg' qXRLcXXC:,lofi\2>ޗ䩟=N)-!|PK} ;,oRSyBuQVd=ʛ1 Rhp=\% S’HH{< dD%"jFd>ZE@H1S\P &x#32x'DPR]k٬U:ӅqƮPt҅ E٦2>^MRuDoc. 7_5K_Lp 日j;:y"HE/9XqA&%SmXk+SM`t-2DC AHBGG^Rk s`1ZBg :;]"X/?{ܶ?%CU=w뺬7akcdHJu I$DHَD==ӿnwbw/a {C#-B#86p jIQjjv`q/}ؚt}eK=\an|LU#u<pN\pRQA '~Ts㈸'-9Vj Sl8'X?G%|aut@%UVH<F1'Tn(d0y=+mhʍR3DPy6KfW"H M33(kD%;qĕNZ`;N~-dg PᴀztRuId'דdf *EEד*]9[mX#%6Emܢs'*cm~0c>śp5d #dj20I}SM"e _ MȾ '\iza Ux/ O[7?,Ϩ(j;ܠ8(  ^dE t8 NtRYsnɪ僩25?DJ\>j8u{ȵMpt $fiTȀ.&mF0XO9<0|ze߮bBș̋On0m0smzyk曽bMdQ׭YMĥ>BWڪה7lM˨,<}xrEtQi 1AUo[˙!K{k*|CѷߖZB%ĊD2,`3"XD&+=5 ^t 1)vb[3%Jnck42#^XZ-v^Hi(vH&'3Wt3V)4FƐ:P8USVEcrl{'X9QSnMȡ (Swcf22A #Qpt [4JòrUcIW5M`81HGY&h_900'xHSD`nk RXqځ)O&x^RC 6lM@ 6הn)ذV\^ =919øu\q=vޟL_N 8ɕj4r88grȱ],~b:X8}p%C U>A}3k0FQ4j*S[-wWTL{nI4D{R5L8`1waʽ3ø)ayU!2-wz5=vXJfOA ?cuϏx~^]1^}"Jäuz;~Gfr Ge @VNq02gD_0Qo,FvKg>ʨt_f)*ͥh˒!J,ȅzrJ5( ),=Yx_m.OSRvo}Ǿ/Otouv>`&5#vrtV#skŶ!^+k)L@yMDz%a𨢚+&9NxQ 9^~8ߘj`p8osL'2a]6zA 3in Lއc̖j84eۓ-{QL U( E 2nQ" D@H=$Ղv#(Pʍ"VFudXDcZ)"V_"D4oZ[ 謩@2S6B%MElmA'Bqw*zT땑ILn~rŬ'H*3LedȩhBw+xDYtPݨg=mK8Q)7Ai@.ld]|3ں.5o*"/Xl^m}8|;E9QIXx]>Ӥ{Qs125q9ALG "k7i#\ F1ZleF͝QFGʁ i}tƆ ^9N>Z|uE&=XnSU'We٦eZJ9\i2ٓ9vM9d6"{M(&3 |ʆpΜ6YE]&8*i1 ēe-:^[Ƣc\8ma5TU\0s,Gqu.0SuAׁ)°4~Jc&Vy#jҹ*i6@`-_*LYyV>?,8'ff%3`20f/߾{.@0ʶÝ"E/EQ47 ._v=\㿼yu֣V0#)6gwrm|S,N%l+⦙s 08Ŝ9s˭SV AeiUK@rFϵ,wJÃ&S+XwEjvY/`2J2a@QE/v+c."*J"q^hb΢@zn{?+{/T迌c00ƌԯ]LqgӟݟFKGi87ύ& !,o5^|B|]&GrI@!twCrC6J d &vJ-ӆbJ E^;1=E*̩쥒S۝ ox h(x\;aCIR=exb$xc8nRRze46;:ƾK5Db|O>4O\n|tߟ޼ڍZgjgr'*uuNG Z\>L?jJgAQP'g֮YXɝ9wZpk'Y] ],/lu)Iq/('+e/Sbq_+B8^eC;|*_Iχh@e Sv!+݆5~^;;z$$ˇ$Y-2JG֔x5:-{8RόOn >Q4g'0H"W=|kc&4e R{eW˪~}R+n?8/dIM?36JrHޣĘP,ήξB԰J>lO9{sJ,HۋCJSR; u7o)o]`]"S$2=M@e 7Q# XKh~eF8XkbW̼ Q<ľ2eq'jQh'˪^jڰ6WB/"gm/NV.3TPKx9\h ) jd IBզhyk xb\hY*|WhI>eϛ墄1NTrI5SL/^T)-Miq]jX;lN V|iZ!x4"Z 06:y>bÕTeb#A›|' .Z ۤ5>YF *"d>b[$ɗ]3AmYl9S.DI >Ev;ޓ[x\EJ`8)dy>WZ!J3wrɜ!h0 h}l+ &qNfc`i|JG\K6Nޟ `c`.?u~G2SVfZ @ȞXy/Nį;x-NݜMà"܏25x5|瞈stZ|"s:P&\9ԁXD{ʹHbY\y6N%ZŅ,yh7t,vrɜ0+ : I[>ݝ=c.13m물$ L]1 r2ǽRJ6v/vK!rAR˧8.wxn OJԚِg/8Q"$:GKJ4&r㈟>ޟ2E#$Ő IILM%&XIVl^,v5F\gu\>51Y1X*gmq4 EfN)}v2geksL cD*9w }TUFL5t8?Y¸BLlY^qZO^ϷUEUd7I:y;mB}GYeemx留xGxRઑ,ң1!{Plc0Nd@Hf[Zcx Su"׃EI sL CvZ5M`^!hӫc.kc`!}|YC/gv{a6  93ؙٗEW[d)l'Y?ldn6)vFI,ɭ*օ_ QJpH,Ór&~#-KoB> /6bo0^\(KLFYaLa^9'@zrRYY4lc_75eLvX `J#;hϾ8R jg ʹ.> A㉀w}T+(eS#ap;^KZ#ϔTb|NJby],ǁq[^[1Cp2b릲T@70To:I#<Z^͉|j{Z^]qpUCĤv2TAb0wbLg#md(2D\{"qˆH /qul"1yG `EkLD~8ڛ 31o=W<) $fEqR+Q_zoktdߋO@pAJie)(VD, y"4c9v+gB Rn7`O7p@k+f8ta h%pI AK7ss;iv3%KLmޠ,oR`<ƘBv̨Pe%_ewHf$+ذMf\ CN{)c̽4rDžhABOA 8:;^$xZd9d{;Bv >ӋFF1Z[F9?_Kl=mi"kxc4l^2Z4wGF[RRy6jn6h}4ŸlTүL.d,dʬAo-o,)KO7*q8k}Q]s۳mmk/PYC\aCXژဩ;avL9sbL댇$53jȆZ8H\k87/+BysX ƆײҸ%`ì#*zY-;u[Q*S2M815|;IY,6Dd$dE\@֍8ϙѩ >}])6f08 ;ˇ-/B{ `F1̗A=؇0S?00.dĀN|X^IX>:ROM3z{B:1(90֏˟,8t&_8pd1I3NR&ts2͇ULYi9ƬR! .=-1-W|  8ho9W:tPٲhtHDžlL~",jE/4R] "H~t"VBr<Q,\QM76j*ph,j>!qbFqLXyBC ҙG :I;ȤِG2 -7riA|1TH3i0hȉ/Y?|uc;P0CN09rYزYepYa18Yìù=y|ٕY:ns<^,K3ozh}/^jI//SW4'aTbRJPX[~_N/͗P*׊=N L3ꅋ~XwHŚmc<ӂ3nd,$FqOL@I%EEy,%"HίջY>K8jSi]E?.B~߮6{)gV'o\}kØ8h"A*bZ f9V^ R4"Qyӄ GձDbL.1j^Ū ?SQ{1?o.?ıXbach{ՠLc: (%IBbNm!&1&eL)r2gCcB0-͇,pr6$arDŁC3^v Po;wb@3tb/>Zŝ x( 9IV gxAFY$ Cϰ!Z~cVPlPfudZ6RVp<]IzO spF9%tX@ PCĬBJq"jU]%(wWƂ:H^_- 38I"0t2CC0ˎ%^̌s0k)F ]vg}:2f4KzŅYE8ocbhn*Cض((,W<@WUV&iF QFV4կf4Qz 8l!pZBE%,6$qta~;*ݏfsOs74WmjEe/2_ܗ!ڻ9By[̈!'>7)}׳|w5iӹWIsFtZ~LID scycw&r3[?noKwiۊb` /o|#0 zt_쟹Y#!Jd9WZ_^rRKʅ~%RJzYw?hkQ<"oϑM"X*@𡔨c97*KăI1/7fT(w c(/LH!yNI ` ->@P_sIČjbP^cVq %Xvv3V`n#c 'K!I7%2Xȍ/yV7Sdj(=: Y߆! ČJ[I(9D 4d@52Cō/Y좿s'FXIJ0ÀI4-XzH'Di 9+x$.)Iy׍ Vea˶C>MM!A7t2;াp^gap!Ԗs>!EOZ,|bbhFweفNx`cT(ۄeuŗCJ.H㤽Ee!1L 8 A ,UQUaV@ԇc.b邢B] d~W6Jeo~Sb6%7G=r淼URuHoe-,-Hϧ]!,AܧCSWr~:[d+k4ܔ02;߽krc$rB6?M_5m[E @Q?OIhWgŃ !u~˽mVKPRrgb1u>OB"Śյ6qjvHș֤GB!5W*Wlʫ>oχ=ubU~LN "P4Z^t#(@@>qhCyR*O7[ Ƽ;Q3BąPd4OKD~xgSp5ϊ ZuIz@M򥱎MvJ ڶ`L *:ԠX_-/LR"$f &/z?mhµfԶV<ƫ{IhNSV A#vCv)(G)kq>fJ$I6P4j(a= Dp!;IMOhUȇIy e_6̰lx& ^ר ]|L-?oFT&Ҙ.I:g4E2eHIO5x:-|Y#:-y:E_ë́gP+2ƺBĥT)d>ܮdV%.+}s% H*j^[Ώv8-P,Ձ,Wuΐ?Ee*Ϫ3vPj{ Y'! ot]Xt:]k/׫kޕb\|{yZ^7lr77uf1>&?n_Om!z\E夀KbZ0SQvHI5<Ө.3<|d#ENbu'Ki\Fʞ|1+7QdqMR W̋\ [6| 6DOEE" hO=O>`31HWYOeψb;le6{lpnt:1,$Ԅ<Ł4ԛBe['d>L(;9挋ЛBPսK7 s]8P4õoz QGtՑs~񠞣 OiI 7r<@-HX{#nދ%,kE`8ȯ?ta ws1c&m!iY6NS|bd&q]cbM+ٴ lme)ӿ~"l٦$RzT0("Ptf1Eۿ׹^rDzYX&|fYϖ[N'uW@hqG^ =Fc;Gќ:3S5OJ)^MljdC* /hPI̭N ⱽ<(XTFA )̍2-ÈqGOJQ :륀BPzFͥH4W&FΚ:ĈdY&8'RL^B"NܼtIW 7$VםA:jP/)2H?aƴFQ8_81ɰGĤUV;_Vp z,zVAdUYpTn'/.juNdBI"ZwX(2gƞuq[/Q)Z#QC_U}4?t`pRJoAuL`v֙pg,EuQ,tڠ*ENa}b[M],쬺yeoQcHT))pG TG^#e=\w,sm I|fhr]%]U-/~NfA$,ͥFD_JݟDKIqd?IH|3̣@\ tZR!VK`q,%uD:{˝U5p. L>5w{֦u|=L^oH?s\s(vC35-:%ț' /Y,e^Zs}D{{lKJ.L~>G?*ѧ T~ LFF?e%kn|2NM _狀"_a0]>K$]NQiȸ uI3Zzv38IWE|^RuOQlӎ<yӛk"R]ׇ]n~l'Gpxʟӊ?ъ 벘`hT{pJ%~M[o$Eh:.i#I'}}1r~d@A{JcRP5o 'h<]̗k|%.Q' n"52uFJd}oeI97˩/67@ɿ0)]J Sޓ)'ہhښ_a9L=F5Kk>_?~e-Eol8pq%k|E<9OHf%?yÃZ=gcP*ھ^g4C,Z>H0ZRtNo@rJU1Ssz_my/ g`mq:<$&Hxho}$ř n5w,YϲZNƽ֩y*PEѫ)%q,L|7]OKڂ`xx`cśgR1&A-XUj>N,}/Fӯ|PS󾚗hpTWW/IJҗkѥSav=lApڵ5ri{Y:JJ4O灅5K c=o}Xt;i&H1o i|VOү4DZ1Jx!6~ߙFH5jyRX_Pn]-&BT+[ ĩy-|rI>re7hq z|}5﹃@-OUnhDOlUc8hZ,~\X{2o[[*H'gWʰP$/N16o-݄OU4FHmōh\Zd,tSx߽>o[5{Νkuj j] [o/Y|2-]<1<<>y24|Ct\IDk!P-z/nO!:*1o_0>/ +n=$fPCϨStWש[@wmSFe^&5ɢȩcbh#a8IMZLc הG!D/H97gaZm=YCA &@ҴK'mef47F H""sNQdrqX+Q'-\mz7;DVBtk p gMtߔˢ \`\-h(-UqLM/ƣZEpNNpΜ+BةSL(FcUhE9dj}j$Đ=Ь9׻9s0XHQ>ܲ;ҳmmqw*:bJB]oTًGQScHs/bWo,_wǨڛTU@BD(H* ԮcVc ua2E<9c.Vq(.sT/PEIŔq=|`th`|-=nj=,a#DG91v4q)"dwZ]{m |9J ttȑB{dC,u3[ O(!:/ )(^U} 1tkdGO/( WzeQ` xh \0CʹHya!ROI nf.Lٔs;ђw#Vvkd3CY`,eK[#)}  Ҷ?4n+0q"3-6uK;7P9ښwѪbC|]?>&iG'.}|T4\`3$4OƃDQHkL R[V?c>*!r4[G"4x;LCweHdͬg1F䑡qJ]?p8?RF &t'^UVTmd&! 52nB%?nnwfH+[r$ aOqס.~9s, r{i"2q`SAq/i"ChPg,ZGI;!V/ldI`yXBxyo ΈpkDVk YAP8Q֝8iuܱ^2u%(x0FLDKZ*LT2xƻeJ6]n(LQ#_ FQnLw=!8]؟(VQP@V?"%iӞ:Pc]O]{oe899&^Y ,-V ,P1u(f˜d%&FPkҝsr&O#j<ۉ] Q3W)cI$"X,YZ}gAT.)#RmPN/WQleOLD,!HZ8H/tZ[ԩ% Kccli3ٟ _|g-O`DDRPL+N娸Ȭ <(`9L״$$V#SnHt:P4FIhވt|3+V0^\+RiaJ?0 7ХB`z-%PVv蝔Bkre}-.dmΥ n`o,%hm"QդyE !llR9# tV{5.Q,&/ b-vNmc+礈 qpy1WH4)5P|F8@uwjv51e 0mFVm_egI@ASAm)F]+;FV'u /bom?]OK>C}E5&m 4 :cCh_:w@۵aKB)lb^LqGl+;YMNM:b`Q#߆B1,xDx1-34'T}QfqVN\~zfFkb1hUq4i) Sܺ%n7:(]ټNG"g4w$aqcl qB\9]݆rSr :Kr{Ɗir WLw_E5b#H=4>]4|c !dmO R/hGdp֝q~duuFkRZ;pb[96YOTl1]+o$y=WoM{hD`{h iky 1*;v'}o\v޽ŕu]} k^"D'hF< K00F:^aU/`3i:TRiy>4?+sGvWye@L 1EI_167}hUY Q2h&K 񤬆Ǩh =;ˢ"UQY_"ۋwf"v{o|7,ݰ;ڮ䗐D[wzm˹RFШe^A+4g~7ߪ7e+쿔 bZ4Nj°B` aՆvHs:mN:tvCmTuok6[smtQ7$_{z!ףd?~VUGtHR^&MMCʨ}ur}fѷ0UW#`5?nf:UlJ#QHm-͵',Sn.h쀮BTkS.9M ZA'@&hEfuaMAO?C5W&*KR<#$ $%M2Jnw\Шmtm p䢼i}F0aO˃rO(b JQUP+p<]0ƕf = <7 _^L'U5PxR tX}$c|@2D*o?|EU]g°}@%!g ޲1iyz6:]dW1G\6*ɒ*⻭ɬ'hx5>*J22;iH&4-,-b,2,RܭSe;V)񷳛Qc3PG:'Oٵ2]+Q={\hM*wwAԺO{~L+Rj08Rb䦭p<"c!H،~___Ω`8;un$ qk'8~sJ箞gf -wTG uOp~2 ?T;fR]4@Gk!êkgH1,g8-ߚp3%=kv38'|U}댳vli-0^+sڈ;*^Ŀ*SoT|alZݣh_B;ݭ>Rg@)~Q+҂`;$,‡{r1hz]qssWư܎b<1~xNdsTR‡C&` FCDU  XJBÎ;Qсo/gwaǗo%+O7V׍ji%ٚ9 oyܩS;r.m>u~F3c֠C9B_fZVI:~@*듿䯲>Ot8!,I\sPl:ƅ8rf2+DbdyNSZGLgskP gު6d8ΧPBg)JOޠ f !&YVaLnD^$)޷{v\O}61%<nR ,x\GV?5PCahm"n$7h@2ݝ0`GٶǏ6!S6Lu^pIL<\+x=]XfTxx$ɮت=0uˬx9ϖxc۲-lqT(f%c1Τ8QaVB-ƻs֒~nsۆY"rE.a,,Nq ֊\&fy6- eT*ґY+m@k8}ZpLGT5+HDd 8WȉɨsSC"WC7]G'Pu>ԄݴN =]`y%Χͥ'~2"3SK 8\S0/%ӦJgwJEׄ(YiU#/nl0TYHBQM\h'w$JB.3\]]m`B`= ~*,AX=.(Ֆ4cDҔ4AxsRJQ+O(ؠoY\ ݍ+=Kb(5t>v[ƌ]%|nQTqb]5@bxIHamݫjAcB^߀ntYOyo$#ke(0{|DV<{;a(K9)PnyRly KśNn<2URƲрڞR7%ZQm})\:P.%$N941)!٩kfnKsQƦw\&vN 0!dX ѝ d!2O῟r53ttUjU]‘,] w$uDJmG4TvRB>\  QnVzWJ n5w- Y*2CЌX!0e g족-ȧi 18%<Ԯ"mD[l XApM "2"M~|hwEj%*͉172+SN),K4ּ]6:R[WO|i;;mEX+2i&dR՜32sPB`娑Q}/z@LSbcHnsP`l0Va3#Rr75~ހʓg{HB9(Bpw4Nqw=*[|XrT^;[-F֥G0OY8_ԇvY<{ ?.ZTN "Km L,&F Oy߹P1)2r'Ԝ0ngvgfz"$N7gvUQW "O?d+R'0&@"w5R\4Ǎ>) _' A'IQ0.RVߢYBDl7L`jTHn[ֵ(x2 g­jXW P"KSaEqm&SefIʴE?*`5n^Tͽ>MqL+x23Ks,"/ Z>a)g߱:uܥVBJ`ܪlش}Ei -gVkjȽiUM+i'<ɆĻM$d0*_Z=}iY 14?6޽˃'O?-?W xioxT&,O@?y2.l Ѡr׃|1:?x`,toJ̄E&s׏!k)Y)fm?Ά8hZ_O~Zn >ޥjdYV f˛Jc! m+Sݶrd  >WMGN0edLjzյ^];J!ˬJOu}kic.h炵|DdB zz?5-N,͙i:u3IMz26e ;ֻ"ݸ &^uiN=@rJB3;/RS iRL,O3+{Ac)rjR| ("܍քC1HD^+B·øʪV{pV "LQFmɧ}h|*ڧz*ZF N3$2!MAJ@A&3!Mè Q %binpJthsKe@vC|^>f,sEz-8涞Tps+dF.K2_kqZH ĴDX)nNpɾĻd?wU@'_ Жp"gBЙ^w +G_Z9# N<adkm#E`1Л.wb]lŲdK1}[IVIT,%K$Sůxx' gJjt{kVjm6aQFkjt>L@OBMG C"ſ; htݍ JS\Ji%@YQ$yA5Y6$%ͳѤeHQM/nfu#o}cDحߐ>IightT<4„] wKƎ]wCYa k^) K=0ڐ1AHXKtX}S8;G븉jݓN㜀5K*8ĖZXZYdhWqͿRt&z'EΙ̽KMoC%ժ6ⶔHs 7S\L5_`]h9E=|566(&*kPq|I,29j탐EVJhdox>ׂ ,iao*J0T9rds#ZЂ0nQg ^wӂOh) W_MԭRP='آT\Q2V WeɭlJ[ h:.nG^ܔ }0b:WWOYA3pkiBV\jC|F cya/\D>^8.ufKQ/R+>X.~x2wL.ClU!4Hp;#,=T\h"\^, |nB{:wP{ur\?_{l/Cќg+4%FiƸj-U]^J|zsTտ]"q?h^SMzKف|/`p87VƀI5|EJ1-J[eqM.T =庞[X~KY~z7W~4ϱBX;oL YpUF/W׆yZRk=]#zw"v?*r M+Ta Ir\RJVeZ0hגF0͝TuG6g=~pw8+5iG^Ė~QmцWJ),[.l Oé]~u0A꼂GU[3 .,&HU e.'+j9ꬴЦ򡮀}=~2}\-n68ȩ#8^qNCJwhP2;XM^KaMHds@VYWLe#L8| kUQ@JV~&VgHUCWR6z3jH .,}?hmɰreAXy~̗'Sɱ=ze$7F'ưF'EQVwiBde EiL wkƀ R*#yQrwGٸ| y}{sk%8Mot>OuĆ0?Ϧq^ڻAFRM߀㽲rtks7.TVYap2-'4qbc d$p^DnDЃ)*[{ ?gPN4GO9KAM*dݗ6.FQc(1r8]y] 242hTg9j]YZMDj=*P![5`OʿAF_Fo;UDūFPOs -"*pm/Q2\jD[~Nu3_X/x&yc7*&?v^f 2zn g m*xkz7%?\Ww\li-kX`\W'qV_-&#p|vG=B:aP-~݊xGkЭ3!Ϛ쇎\%Vd!"5Ie4VF- \7#XcU:2cY4"Tx"Ijg4<,/gxGe>=ǰKm:T,<.sa/ؕwGPU7&*它)ly &yJ%SO$ۯOl$,c$ ttOQܔT5N: 2p89C5oIʍ[~l-}@뛛@Sx]NT;jiXU,c5>մN!bc.qݳub‰ DvKCvXy3 ܲfE"E%u\WtLn#>+SP6C*r#2,\vE"eZşYC -xV*b:p^B£(Lo[z].hhYWvch^ٖ*døS1-߾sv>xXcaaxZKpNF+zJeO*siR]ɖk"[>18>m?, g&.1_^')kL~ sy9~PLj߮o`LEbdj70!˸fimtf+Ljeގt(!\gҼՏn&:Xg0iN8u>/q1r֘ѲKSˉ3N|9%^6:Jch(`5VԨIu|A9o"O}bڑu#$`80L@FA G0ȖK;Bk-Ԣݎ \"!鐺B&< \B1S nѣDjGLX fٸnl@AjN>AAFѤPwֻ) ޗd+l#glAX*![MEG/պ3 mAgLD{f8EanVl^˂QiC(ȪyIlCR1;k3ͱZݫa,Z $7+7}$0 +Utq9*}B~ MMFl~)zKhChK-LSK/nUH 3YԹi^U{`ϕnx$")ZoJO9oHLzAWve;ן`(5qð95b/h͟!3ͨn}^ʎcJ %fn{w+4VGFQ%!ezze2Zk&Xz4x4 f4eTS}m2/?Ӈlm?8Jţ~C&WT +v7 ĘIEIxu7*w|js 64]vR 99Op^F.vE FShF՞Ψ$nuɋw3(a:fTC ؒBj|Y4 2z5z&Gel^& 5AF_&|ljmO} ^Xnw<` ,Y~y#ߦ@Q~4Ėw/;wi{iJ)fJei|:`CLg!`)rA,*ϪeWEU(W9- Q Vv6,OA5\w3gi< 84n1(l/̅*n8 | uϿn }Z1M G\SjRQE 3Z .uj9ƛGHR^C_SbK4BZFlj]!V]G$;دw0#gI3oA;4"B;l- ds+^C}dx PSuQ$E96״WcDk)jbWۀC5mA{6u`J(b-9?djWc{e[vd#DcNYJ49}ܬ8^H (% M]]Q+2*W#hͰ,v®sf]E+(w]QE3DZV^ )!%z(n.(w/W#Ky^OrV@F{=6"vhG=tu״ވ.bFRoA* QƩ:M3 gPU*HX3O4 pzM>J(#Ii βk7I), Rw= x2~?z*~ d_--W^ގa?_>|,!p,53Ea,CdE>gz"_A.*W0s蟮@\>ceAgLU&:nd1h3嬡S,?J+FK)0esqYMjRMZd"TL;DkLnr5#ِKA,K?rS!L:%H4 lqב6flG(#x\M&σ.tqVCc6^ise@`vV;O.Gg t:xM׵AOX4V&j^LJ ޟQp X=k5*.,q]*]`Y7M!Qh*Oym+ .((ry$P0 kǤIxdu)]-.= Upl`[캷*sPd^ĊR(QNE(d=DA`2W>2ԕaY 8icdJLB'^qx;kdYT۟4"F!ЍoǏTE m>? 'V셪xotLnȼrhnZV5Q[LPR`,K,KGƍ"C߀ V5f(]dcN%"7o`j.ʢ38:JwB~OdWg:sgHCR Vh.Zj`~1{91[/Yrm y?Gf.<繓=Ҍیt77-U{zr-)/bn\(ӈF6o_Q<ޛ糽o{Nj9ym3YWG6N_F s\ɭA Зo'3ԁ|E0av:'y6NfXg"~-sAl.AÕ&s̎ث?'nZ)9%*i MUM4ihkl @eP8]p!bI$ lݬ lgиٜ-[C Y2[+P1 "06Y5+֢i5V4^Yj)xԟ>`=W㚇*Ŧ-U쮪gwU]̂PGojW8ge셗qxZ6\Ky.6'Rcv#r)R:w=bH v$JZ3Dޢ[۽ƑR?!~ZFlQcT0="UB(v,cu9 p"a0T$hV/K泗a>xU޲»qddY:t|Ƃ&k2^ER{[ iOl2+B4-ZƇoM5"c`?J*%6?kG`JddZqSgg.Y,) I!S:dC#zDSXVw_g<&W&!U9ڞeCh Kqv_ Vx|]5gE*}5'dR/ x;ʡ znCut# Ykv0PX)4AB;p Us(R1=뾚)sN5A P'hEBz6 w<<&c M1gJ(vUH, 0\Sx*œ`( )5U3hRv&Q sov09ײCV?\Q|CZQ-v4cîsq=c'oᴾ5:S=stv]tz[zYdӎW9i\-V< dK;RvEByEd{"x!Eo.nx^xgPx~/PSq7KN˯z,IyӘ8?tY kఃޟoy{=ps??c|+\Vor磏3m*Hoo|ycwrgOĸYqh~wNBD ۶Px6 /$ws3o$yLs^Z*sX: 2F3*:_Ult͜0 #T!2TYS̞e\ÊG5܍^\F,Z YE%agAUQz|4sJLl!yL@VZ Y gD01%m:E,~<Rؑ#]d]ȪT4,(0  ŶaSPo*(5jc*Cr b2&m׾YU\Sx/GXDu[ܩ3c`::qW͸+d/6"z'+n%y&oG"gWv @>pqVI]A:n?x- ±wS)V0&ʲ*YIe%;mR\4^{1VޱU‘Kƚ%v|:3 ȫjTIB˺,j% EϨI7ofҡl:)h[Inx&uFfJ$p>Y D7,O^kwOIS@ЭPuZX &F6t퓵={pU2b((9ezJFr1dmS95ꭴvPLh[#*#Td {ʤP#E N-pXgZ)49`K Z=Z׵e(#le< 3O%"M4lE*-i+l*v2ѐMu ik_ q*Pa; Yml4!";3Ph9xcKpS'dv5r5Xyrr^2,zK㛯.߿?מ2t ?%˘MUC.{qԂ ̆UN#(%xجHr3bgO% aO97lDd̦{`8$SbN[Anx7w%q%9u!XHiQ̭,WI\ar3R,?Joc2AGHPIRYߗ3.7 'fӤNQiC-5vZ_ڢ=R$rS]t=e1j]|Zt yFv.)@06A"901}, GWYjG?4inՖ3!\ű@nly&CVZ]K89EAi*T3u +VAw-P Mq.5kJɂY.I[ɢvmW5mT3כ#dq͊yb^;~u҇V4կa9>L3:TQm' dqMAgm[~o"QB#8>T \yf2Ō"?PSSL,~pbq9Q@Z>5SLYSC2iS {q2%=|3)m"T-޿#qc4[`*e}| ־ZmD ͏&>S瓶?,;|r+Gmb:A9;dTqOȗ 87đKDs*8@+Uyx-[oWxn -j+0r2_ڹ@"0߫Dӱ).h^K}cbCW2w ǰ]n.Ɔ3)&(f)lCvtU 1u̎sZCKe4!9]p]D( &;mE9D;IkꭐpD@mPhSJ zM ^VkhF?Ej^#qK8$ʬ5usqufw(zQ{ dPR"P߉jw!i^EOTJKRI5fMjs-g,=T'DRg`)C0 l}",J,u)9'|u ؍ռfIMji,$:Be:3xEgf]\eP/gmJy ?\rWRm?`r |~UFߎ2`{ᐣvoFl7go3 MY a-!L&uKk[ƒdGA֞%-w+Qw  LV`˾ĥ]`q]?J@IrcP` T߳_L uEΥ)7mE1h0I]_OF  F!#Aj1)w,:!ad[))+ဥc4Q/cᖋC^5y^Wd[DPr4 (95x%&\*TIԌrA,UJ0JC"[{2P~TY@i6g22 qͧK+2@\~=.wױ YUJqBCI7=`i?KoO); :7kWkeFPceZ=Gހ} %-t5*D*.]ҳ)VԧT.I.Th$mˈ~ U[}#k5> c\GJcHB:ԃ(i-`A͘##ICrADCv:W%ٽv-?_e"LhmEz!ׇJȕo%/Ot=ǽa7R 'XVŹ"jeuDK5~sv'Ӑ't(y P;3`$d6ќd'wG1+kS#5ώF$]1 NHHRi*Om(U{P6rؒ퀖$&Oч=TI #sUbS( fwT轚bSvwBK=+Y+ARR_jUޘ}p;,"IҍelL-^u4߽Jc7lePtuӔ y=diٴ5z]Ʋ\-cY=&Qǚ/VtإP3-nwD=Bꘂ@ NS'YT ,E+JR5aj)HwLWi7I[YAE}94ZH,=i\/q.Ny&)-ϝD UXPN8e.Nٴ 󙬍}$w#(V2f9}.slY4PqPEVĆ}E#ɍhc(vיzI:S/`G<>j>dqgܷ{6|ny1;>ű6V)˾[>\;w%q$u!̄ +=zihT;/oV} UVXF.RL*-bg՜L橨r&FK&:5[k-%B%ᄺw7vWW.їWg9; wkB 5dIWI5%o fQ3SΤwuXY` ɞ<-Kt%GA9]+ ;-%\`EİoHPl`5N=q&E#hJTЁ>vh_VQֹ2NvÝu]n<.SǙY`qe]'G4 ǵHS _w5|y;)*?hM7^×@ 9tQQ}q4X fyL]HJ樂dzC􂲣.Xk,z5C{a)A(:F[fKxz8Y|74θ=qtt@l _vj{Bޖx NDDiSS~.g8f[GWBɣ6Yd϶mَv1 rgT^KPЕW[|.IgAKWHtW\a*A6I :$[)A=r'_Oko-p:)dQ^%Z8} 1:;ݿN7 _N78q"t;L_UYbżscMZM ZHL.FZj[xϱѴn3+|H< Ʉ[P*Gֵ$9FX[ձd٫M>}lן\-}#.9cFՖZj459`Fijb #!HӔP>= /qH(9JJ m !wR,WsFX8mt7$}lw_E/@ux'G}7B aL!lZ刜dɎ'Qq4r)K.rw MME"F AIȟOG}*V?ݑu\1 JhL*zu屖0Y ^P-NY:~uhAMI+!{P6w ]ź RYpKކNO|<<O/H-L/U 2ўZl-yKСGcnsP]zUCՅϣ/lɠ%Q˘)//b"EL؁G8R:ջxfEX1j+:p8Hr➀@х%52uM{V.apmsnz/nElgmS.&S7_*IqCnb<~#9*LjWJ\N ư-dT!(n}O.;ennR\.'e7 ͟۽T97~v+comٓ5=)vkz0Pf"4vXkzޚ7eVZʃm<߉F V!1; NB8YԮ%O<).@jJ<8ܓ>mH]bLAS8OVFI"+oV2[rmVJ,|tiCQ7ZCLKx$ g>뛳KGL-jMey ]i $10b4bINZ%MYxaHxnsL2J:\ ߁6E\ 3T*7]*ΓRfǦEAT դlDҩ%2e'[ fSBM%ۻ>0*a(̟>P税EMT`-hPa0>밢,p]eU8*2!.5\ c]!ڼnG ݶpR65: fS2k4)$u6K>mQkqTQذG"Sj8R:.G"DŽY*/>yU!ܻk& 4z3 uM`[TӒ=#jܚ& 7 r/)M0#4+kK~+<[?dhTEXzlX 1p$K-IR> e 1yLGy9HƼo/Q+k%NTԥa)rX+j(Zy>(Ы1ćR.DJI*>Zb$D҃i#Vca d HP$'^Y.>}=/RK4_?=:qyF֢8AӬz)BPUT@bK`{j)6אI~1ul1ʊX֢dN(1Cꚇ{*X 3-5߲dMi \c\j!E,i%>WQ z}(.=^"ySݲ0%gݨ!zqmqϵ4P`POE|Er<`@X-p+H 6(8p^[{6jǞͷwn[W3Ynh8ZtD0TJ0rNժAURg'}"*8>ے I9uaxֻu6 IJ{Ǣ跑t ucØomC~d(D=Upjd ĭy]S/ףGzAKJ'X JO{aD~39 9JpmiG.Tf?$HZ/qOػڿ}{<,OѻlՊun /(wO 5ʜIEgijQYCH CePr.ݥz#Q6gg0>UPnT' ѯUr E_}ߛh|A<ߓgHٛjǑĊBF\wa qVI-zSk6(w,wI%4&jhU'z+Sxx(utlgp?˶˹=r~ѥƹS[8F])؅w_f=rt1/3f+j"+]NpD |ҹ4HD}SNUT)Nǁ (Ze. X(H{M)hl2qe1v-2A@Pb9"M9\r H3XUHtfԅtԣϟǸ~<oY }?ftTucQX0},]Ue7=] յ[|Svۧw|cvy%zXxOϔe+l/]f!.Q3=8~-us?38S{/;@\3~񌩀N^R(Ĵ&L S,9衣`vH\߭1qW أ7)b3ۦ*YTc%"F{lSu0rC恵+-~ D$q *M?m \EfA\iʈ=]#` L]x#[[;=IƳc>pػ8n$vIփ,A-p9/ ezExHjizF=3HfG3T7YU,1/> x!q/)A;12%Nh2wkD &xDƼԀ J(?֐kJ=bńiYB$Ӳn8piY--G>n7vʝ ɫ 8qL$7c^qIdzyq ەDsKEk:1"Z\`5c`d{}`%~^ JSx}`]y9: S^[B< $~dy7KM<؄QH.Xˡ'8n98DA1t]xesԠ,"1d+7y?y%AE2:v]y1 rzlVm:죝oU:9ꫴxhdó3/s:i Gw'Ν8,s=%{,o8Lbe#'ΏZk^ÅTJgIXbkj!5șɰ 1V&_[El/C]vEUdIWL>*{t^}fhD`coIqSy\~*{mxqvgXq*I֚z/פqu{quSv)v"Fi-*#fx۲+ԊA3 Y ʣsap`t@i=͉s>s@.G>~Vy^Wlv^Y H3 ]Rg1 ڭSU~QWV*̕WX#l]{zfWQ6'GB#6m+y0!~0upIaVbuaZUZĕRXa%iڠjUC\ӕj{D[)a"ڗuK`U;jPTU؀[qQ]4+|*a 6̓ZgfICِo<ƪӐ6mN/蟛\9|A%{MqjehJ?ǚ$5yJ PF qjK}JaAB3 _( l6D%Z&16eRmwBD%bm v,:]6 dI lEX"0-+;ŘD:Ң >˪ {)}W/=֞XKlt2kĒ=V('eRLDNgU) XŹmh(ZԿ!Z(ֳ%N*XJM3yϵmrUMڪcN5]_@퐀lͶ_CZ_~pp3#҃1٠lj&B"Cw7읁kйKfo f9gj H'#mYTmjqUj(E=}s =QD]1=X]!p5s<|$l]3.e2fߴ ?+Ui8ۗl(hX=pKg `?t֤+/R,Ale7c^E=R7!adXUn)9Lz21Rv6v {,3!烵လ M4N>Q.ܓul5ٖ(^@v= W\p9:T?M}UQPn,6u=g:` cP#08u~i*PCobdD% +N%r abqex7!;PbMFvى 5!;S@#}:;lr 햩>LHgȟ޽{ؽy4k067(. |s#ДMѶJ *V3M1k/RVb;x,TUPc]?(I 0F-xKacW:xnx/u,~-~ UHb-,snWl!MnHoIB̰a0`[#C; \KsG:sVn}37a{:}G8O Nߏ/{_UM_@b+F0窭 XEv]t = ,<_ ~롕߆VK` 7M4nk][94PGwp C^b͗lox@aYs+HBGdV (4fmƊ1~f)u+RS8-">-D!%EPłwܠEv`A{M;QgE{I4biĜ>Osbl!2<[lhk[$1˞] 9ȝDnxWuYϥ}?F·IZnůU* zmW^//G|j~*/Kv$U4/3ξ YǙ==B/ٌa>3͛xS>$,J`bi(9u|cMЖvu(=ӎA6ʛ.G[ig.T)a0̳| ҁO|5C4l>UB3N-3gHO1uSm:6īZjg >wSbo9@AzƣÓ|wӒH 6MCw#}q񙯚XecIf{ﺓJA}E.}Cpa10E`. TL8j,X'C!Y ʼn`X~)݃/9HܺP5A"TU,[6Tޘdft}ްmݑY_E,ON.WbԂwvyҕ@3Rt1wlsV~-brE;p`NVr>Us . fMLd?LD` {o3_N_/ MףW?v 75{0ɂ5&];}fO-msk`Bn e<{ syWo#D$t/я.X&7T٬Tr`W8ml.0<v:z[6LM|†⪺FG{v_|޳lW_vmSht^v4agd&&L2l78 ]uN0 -؝qU,†+b9nOO03;35p#HRK%^#b7[eT.H k8i1r@|+la7 c9pBdžӍ$|(q ޕ<-Khx N18N"SoۇYˆ7e&Ay$٬d'w֯vQ5p!l ɡ h$x@ = q ])$\xʦd;]8g1_}e:ðcW6Ke+]T>>I_`YZBJDSgo6- ~ՊP|9*N@ uZA? / hx,\7%ݽ#Nқ{P%s#i, PHBEӜP.U4\7&7DE=Vl h*l`0,I> jr$HlqăS=v1K/O;o/anx;3G0[&x7ixBFg<ts?z2=5;TvZŧW6u]m)BC8nURm: ہ7gʔ-15V͒5S[roBg}%+0}ٝR .~5u8fuXDb\I]3^^ КC:708fZ5܃:C\3|W5EƜlt\J+` ßo@Uq:vc n?.-J!~^EGj)GY6Bg&!o^{5D^1 lFHxV62/J3o\qnŨiOx;[s*Ͼ!šg|Wq={Cޛ]F};,Gr/f@. DҿoGxtUNr/ݼ=)co;3 7t/uD͟f0޶F9I{jAJ0)hA;h&|NQNw粮cI 6Fh)ɮ)eu1+1~&lU!EJbWfoX4nG+ f0r-66Ӹ+j!Ӵ/x%7wN w /৸bShrrKKNxFFI%R%3XЂi9H3\i%zhgM\bljޯWq:%K&z6dSw  rL#|R[f=A9? S@ Ђ 5=A-"{*y%3C?gPH);W}ygR!p$e˼FRo#e%`]<"I'p.@`o}rմ9=ُ'dT\KJ 63&!tqY2ClJ!_JXn?~*4i;,\8Q{g6JV(J%.Qgk)$A"s yk6[ƜA80311դVʴ`H t s4+9Jzf?+Y=I9>HytEʣ+R]Qͣk"AF+")aID)TS5j.@PAcf59F+c6`TDBbmӏIIm^)/bj'w|Y*|!b6\ n}YEЈ#4^ "('ʥ|78`)1uD RȴA"fñPiqהkW5trIn}uU J(T-H/4:Z@z^PJ6"P_`0R4uPGL2qBZhH.} ΊBI`rs9'q~6YׅS]H@X%vRcr;lj&Ta$ Lcy66'Y[=Y~JL`̌h8(Qd#RgS 7mС{uV# 7wߝ|:WOuJ9 kZa"0>@(+:I⒩k%R D]s0X5{8Mq vzUK) wz8ݐ3Dp="f~~FW]7d%:Fx_d")R5a% v6Ox w}w 8` em`,Nۅ@. ֑ T#S+oB(\-Ԯ@ˇo-tH6j9 Cȣ(>}T6pbfֶOO?ߛO7יaQ!A,ޯ] tnYzvEo'~UA`*7^MI zG@42P Le+~%<}zo(`~vsFm,m)נUN&IU-'Y̲k\!6ϒ{cΐÿ:Wl,Ii PI >C̎xoh{xe*Fk'xJS h- B5@RSmꇣہgЦ.Enͺ]Mm(״]?z[kN.^6ϴOkK !&Z_ 7CP徯)R!ř /Lv@_FLLIا:.EAWt{wfmo͞7xk6gim5ߵ]=7- N_꡴\On猃|\߂ 9S r a @&j#$14'Y_t;Xfska%y BRFI_IChы˔q`a!`8׹ew&9l Fv!H4D;\k.Zՠlpm ʖ ElZFe>F yk98A%=4A&NJՏ1N_% ujZc_I $Cw.Y]*;M!]Zx "}$ÀUi2'fg^:~F˽=]Q.w_2;:5bPk<`ώ5,MU9ϗw{rL)%fh 9NV\֡Y pG֬o@m)ߛJrm nÛؿь6Jye<NDt+ {:&3vpD$${DETp_X@`%: w~DL `ɝkr WGlTE wEh$Z_z2huJ! ;#Ū),KIsp(`}J W>a4ˎ;{~JS] h ~F_v^7ռnYr%"T xˬ'H"fJ##VD#R’Hf1R:rK (ז)@t;-7 _6B9fHRIwg7yG`nO;3 rx54lCA5 ~hBar]( rNpm {2Ã7@7Thaq6aN8mONxc[Uiµ&]_US;K@REØ/D`b6\ M`U>j<6Gb*Hoj]lwRЏ s]h wlXsru1\vBq)!L*3m(!ʁM<38"1 \(Bb<"jI=Ebm^+,"mG rԺq//Ƴ/%q"s+a3Ϳa(pMjq^(1sGx+렢*PM` vFHQgD((YPjr^BXF >pp}J; Q5(pL*aiܷÍRܬ$>I\ ?Vn76F6RxMK]rIjQ/Y :k'U>Z:ޯxKD ];Cw5KΦ)0csEw<}T41C tmX۸=fP)+-˵O fQLKZ;Y2^#?R!!(\c3Ү h#4ί8ԫp]3(pw 7 /_&?a,~qqimyߟVG`BؗSBuKcݗoo.aX1F:6_n;7gsƫ44gJd˯>oYA1wcM%F*~}x:}"i0qD\e+:8͊uL0h\ Vm~T2ŕZ'\^ӟ -v]*ragJg٫[H3ii-D(֗8ƒ(KxikiRɮ?JnTwU?7`;1젙N8>O{}I%Ţ.ћf886đ./հҀ>wْ &'4"R"|ZMPv&%%yVd̓(u#ዳzJ/p*eJszёE3Նܧ\^)Tu;rվ{FӕtITʤZKqP,qe>N{tqZR*Ȉ9JA[ aK#]\Wb *Wm6ؚKh#lόPke_E* TC3><Ӈg6 >=Uhs L6'v5'v:KR;SiňJQ!qcg}1F/sOU|k x=Y'sj2A\B#d cKksy#/X}rH??͕ O&i'VKTׁK+f1L]pDXSL`W!̇h&3ƴfR[ܽ?oM%vƜw7E>쑶mofg{f|g Âpur{knW>nMMd"L|u` R˕K2p֔XwWW ?~dv}Ϧ>PzS~JTφs/}wmEk-EQt1w:d0 ABB`Zv>^Qp3{Ǝ!$JnB"/ a?'[3,l1ͫ\씯oI,oOKM?7$ï>A' \|L7p/xZIHa_ԣF_ލ>]:UB8­߯R- C^B*Tr×r=|k7Y .vU(ҶnGp([=Zq:ncD-M xc)$\oTDŽ{x{} fg}{2؈/C٘i`b9*pԒPWc޾G?9d.~WM+>}rrd|wQ2vEo8}ZR!k|c&*( ;@Yt !c{p"Ȩ4kE_}\[3餔g.]R*t}ƤgAhB Ih@.0K|QeFWC"\MNJ:Y?$| E u8cŰ^ RыqO4)IduLx.oHڞ½ #%9 MhjPEc;?~PB zA[qy'eArn_'3d:|- 5"o+ReQ 6JF`: J*Fʑ J4aHJM;g džS 3gV6aO_uGJ,=5i<:5|atI!eI>J/oyU)TjRuϊ}l`+{.JJZjdKG: 5[$jPPB@$_F(5RFjr#`urmkt `J' 2"%XAj:WwM_H5}[V+؛lWlQJs,J?]*TLwf^w߂1 iQ܀1SQh2c8 PdX9)9VE%dw_(.LS"/0H`W8F0ލPԲm˕ew0\0xz M " a 20 Oh+]c*D8BȢʥ44Cx%Gbƥpk%&mo}8CM PӍ7C18ŁEzϮ.18q?͈Bo7+۫qyo5xuoӢ$[Oq/cYhVmRI=87 1QY [o\! 3F.cJ4|sc-@WO!a!D:hj$F`IZFe*F&|Ңq&PKEU-v% dcP:do~_aI-U4LDK3H=f^Wu hJ4ܕK0 yҐG) yҐG4w%A6p^EVgeTZI`j)ᖘtR I]5}Q#br6C؄aґ4q~f fCœCc4A`s&D!L*GlZ(MCW*NW5e5IMlgH-XY2(VDS1XG`:,AHEM}ZWhvy̋R$|mS{?= Ԛ @g,W*0:(&с_=T{9?rQh24ǜDo&gC-?@1MU?s=GǢI"뵙R@a@_WƵ1YZ lVզyט;@DQ $ :Tp*nׇ!͙%b-к?lPO ̕hv˳`؟Rx~s+L_37ykM?N]KOB|*6H*aO2ܬ^nT>L^~IIaLIfRe;3bFLS3'İ˖F^$>m9*#o{V%hd7'UϜ: IZI˪$H g~'R`"BrEV#0NhG)|G)|TN:/0VYN csXkV`c#/jt3(]rK $ TZLNUNSs]Q P}^ pJԶvRԶٽ'盧p~n~&5*\ζӰJ}93ȧyJiV| j8 |J.yo𥹩 1x]j6BO{U0Tm_hCWh Trbe+ٖ:8ٖ_ي'[B"5к[ɷ郓q4#*RTb)v)2D>yO[:?{W6忊QL4J6,M2Ŧ7A,:A7vU|%%Y.+7[$=-bH-3E#@ F泊˅ # @'X@9v[Kk  Y%,]+Z! kFT @"Th @DIa0W$Rq$dJӰmrT@L1xg0aZ1(XāVi(&8I47"8.C Q1d  Ph3x,tX.#@9DҤ;eաbx @&a4"6NĢOyo[ Hzt)%Q F}.c /VĻ"VOY=-e^zX'̑wņN r6Jِ0r9s㎼w8WbCxWĆx9lKH,ņp䝰c$X΍wElH!Qʆ|1yr]H,R œpy%% M8h3U [pP +0SٴjkŒ|tu T43 Q` hV΃WtYهϱ҆0zãsyc#nS)! eT5%.𐀔^}WnL 5tm5 k.}/ ǒ+[+bI]Ą$6\-1,=ָlpJ^ce δq6UH;4ҹWpgÁtmΰyu-]ZvZf^YOyI3 @ p)b%R櫉 V" ,RPsl"QI!n k%xe[sQ[\d_63kewign%{sZ3_/\ͼ~zexf- vj| [xSkB]x69eIl]$4F]doFqF52j/ݺBgw|_ ,P _ ƷYCmI!E'. A A]@iUZUԏ\E:PKSW8NH1R&&<+ihƄ̀( `IBD4E\NQUu(JiJ#_s\v-8Bt<%q. A Ĺ4]CͰ^dJ$9ߏBxek.8ڒ3rkKnזVXTq*G! h?+TD,_o]^=bQ.qas$C.HW󉘏MK Ľ9TiLKCFo0$: Z>؂IиLeLn>|F33gIA/I .ڄb~G.[sõ[= ?XV]cC?Y\Éo39b^qbLgzōmd1t&4v?9;Zr*ǜ-x,ngqcx,^bIA9U'$"y`__Au2nzd6n{=8#h{zӜ*Xv3Vֺq)FE8Xk 5mxC֢jdpU#O8]K!BV2Tt` Y#=,$ 4S:VS$ a R#"c Xb" 1 -w{F,^}u<HVOpuhUf;dDF7z|%ſ@u FQ [s Y GU<"+BVyNAVp{__JY%HAU{e؟bP TDKg F q$H).'j.XԒSz)>cw˃ѦT2tƻǎ]H - FRjrcJuha*f˓+b^RA-o}Y2lұ-5&R] A*_XzI#8Ӣf2$=I/2՞iuS.%prR`Wp^92,JIz6{r5FYi@ y.O6zd)}/Yc\m<f)BkF7٧Ljg7kq޷A|Q=fٮojؾt~z ƶReFƜsĶ'u*όk6 ^;AԙD3N0L#_;0_~2H:l^Bj6I^z;qu>ͼI`x hv@ţ:SQ?]?5/_@la/h/|zb|gFgSyv0ve`4Xl~{iޔ߻l#uS_l~N%=wfK?[}6/t0w5nFke6LZ>`6tկ/>N ߘ{6mK ;HS96^~ճ`gqX25d8>է/?E:1n!¼\g_Wdo2O19Э0.n)lk#컹oy.WOȩ^zseˑ beBh4Y$e!|&n89I4: d( Ab@%'T$޷j~Aan#blJkc0I\oɎ}븫m?Qx @$'x-w8-ѧ ޗ<ԖZ)!*,N>!A1 c,-#NVJPalwXx]PsA e8hɷ%Dq˳s Ia`޾C0){zNbe>p!v0aȿGB1%Q9S69ĘV]㒌rh7}Z4ct(' .lTz8sl([_e"[m,z޳]=|2zdneddW'et[3ӝ݋Ѵ~__?vd92asiÛ6`׵Ǟ,j %a2g^L*?JT$(3lIZ x+DŽr֡nbhgngvyNz5g,;AM0)LQ0lG .cp5&TqXkS K,[ˏwm_!ewμ|E-0&nmɕI=C")) e3sΙs~gV1. uH'x@c=q zqM{\ .}9R8H2@NZp =Sul0!=|xCvH2F\4ΣM&ߕILfw) CfvN Sp ka`c Jc4$X2TLXǒ=KcKXd!8 S% ':ՙ)%DYk97,qv%ƒ,d4  8u}\%{,cǒ=dԹ~ߕ:'qڲz/259)VVJn3,;3(Rq1Tb$"e5<^_"ZŰ7E{jY>,]aM;PnJ3Fd:;,,b /~L30G3HNx o: C'avr=N-O\*g4αm΄ep$u\`"HLy'T迗%lMrPsc֦]&fӄ!nvFޕ[(HfŘ rb]ƃ/!G׻hMP_魟M4)|}DUhv23sI`~9kAIWd ̀HĨS#ƒ `VvSIC;-Mj$Mo ˚1uʽ눺FOFԵ "AԝCY7,0Y%rۜf&7FVϽsRk{1 kHs-9'ynE@D{:h~ȱZ֙VD'Lcz.>]u Sv, ;+:*&)ɫ[J9+rTL D1',W"Z!/s|MĄ>32^(Tj:P-tWBð06fXaSe=SX<#DҜQbI] ̂PSa0&:9Ѭo:P#7|o%%z⴪ׯryKvZDkիeT?A酗yef~B0/`Q 2j{ ΔALZ4oJΒMݵfTmY@cu2bmX0׭Q4&ey5|Zt|Ec.}+gä7gdE±7Oՠ.XUP|DCi@Ci@Ci@Ci l֤ͦ $"|bAz`xcf~d? YswYtWϟN}ى=MS=b4&_=>0>1-^~#hOp򑆓4|Փ skF-PB̅g0!c)#A00vRc A@xݞQiB-b6b&a&S`uZx+;/izDz8 ;j'C~I*9&sJ[4( Fi(ʵ"2P$Eւdgk끂,e;xȬmHDA΁M/x,j/<S{̳x* i n{`B;r&D( ZYq1 d'G:W\ ^ !<qR%S3rGÏ#w|K&&d<XX0O䠠#Wؐp} 21 A'>Ϯ_Hn` tbx_㒙KKJ".tm0LF>$w>fcK0Z30 AE"mW~sT?h]j*ȌܛswxmWzN\#rp*FT47U%0WF^`En_3h }ք!J,RP~xoU(DLoG!VQ- O>TyC9›%kuԂStÚRTZCэBC__70+3$<kXIAq{?nCQSAW?E|iV>uk^Hi/ A%J)|2/AJ!QoAWִM{OH @4SӖ)&• mz49j~!6;8e[6ɋ|4T}eFpW{ezI o<) 1/~fj0J Gޭҵ]I9;oKl(9#a]I帇Him [fp|szTzL}y~67d# NN w'δGdԸUcs[m~r5J_Y/tU´Xv}vrq塃3oexVU(T?ڊ1iҬWDGv8İjo 졲#nϞˬޯm}ޡwј8Xg;8a3T)s섚(S$jz $u+>EJy(zKQ;F=|1-N[ubȧcT3ߎFgOۧqQQ֊8(qbAg7@ly'7){Pҁ֯tJPZ/lBS9H1!i A$㌹u5nuF(Tcʨ>"(Y C˓(^p"%_,H, :Qָz 5!':IK=rt*p&R6s>/F$9$J&z3Gɪ(Qs(ksΨDKP ي $hڻJQcb6" +uTJ4Y5k4YFDze" zGZ֕bz/T|3Q:-lBZq4Q2&3@V:2a8QsDL0f`QnӃͿ-<U_]ryY-reS0EV]"Lp\ Pr; &pmȘr54tkn6oyG8~ˢEɝ-[H|r|;;h%]8ҘZ3^{3`⃦c3 /wc?->5 Ёi,m꺘6@l_|itX-Wj H}Aƭ&#ȯt)DѠ7Al:F~D5&ć3D" z2;Ntũ_nڅ̯ hhGGV7|M0uq[8jpq8bVL\Oy!)Ìucka*r1DŽ:ta236blBB1rF1&"dx'y庵]Bwv2F#ʌd΂7(w#RNPAyDJ"tqKmh8^=Er+Z}/z=V8SBlXHqw+*+dN%`JI+Lޫ#"R˹GelkvUƛO40qד-Qًe4FT|&gL(ΧaލA9PÕL ̨"TM }لEM&)dRͰralU)_K4s(n!x{8NhAG7X`L/_!f SFa?"sWRF,kZ!Nf/);7ZXG=:YM&%go=fs[LM 8!M֥)]Jnjsq1k`(( Xߍ4x/b iǂ1߬Ԉ{|F->[ܵ;$^y YJaG)K蓥yN~OTxź#kLPa=P<\}ol}bumH|)2̘8,)8KuX*32ϭب-U]1ڗt{M(?+6j~^}}j`s:Χq:vgC.ӽR^L=vZwIYv)  MkNގ>05{OWowS3r&?^zEZJq?=8p;nԃchu41.\IƶvbE쁘Ti)S9 &T[Xa#6X9 >*YiTL D1',W"Z!/sKĄ>32^(T#Yr^3[j2u7/m7^tB ڬ>|Rc˘DO3o'N#)2Ԣ4GBKDF ?mEchWك+=ioFE :_U :=a'HL>FJ|$(MSEՎ/Qd;7IӜ+2dC_ⲵW4<,CS >M)1qN3f,Td%I&gis~b+>A B6+Oہlq fhR8Z#(P2W)X!cj}*3 RZN ah.܇bh4Ixzpt= ۟]="Sj;SHX~FuFg7C@VF|¾_Ä;ullN3PazDIq+/rui@3Z/YP}&SRt?lA/k"6%S*pdi@-[;lxx9E6uC` e`P$RL,|J&]x]=*YfoyBaWc{8o[]mf.nIoun[PK }רҮrQ%H3A{A-*q 8p3a@sg Mtp؜6aʅ⻶S 7nߘVhyp+FY%ׇ ɻq8鮷䫾G@c]ȶq0c~DL r ,}@qnQ䫕u7/46cDÍ-Y]q2NH5t0b3UP'^Ely#^M!okdP~壼lF 唪c2v#7SA5s+jzLpNcYԿl}d;5Ck_]>VmD).MWI,M&Uw룟Z Zu7|'6.v'x=c \8=VdjG4`OAW&Z?pqѩ}=4.][>xc4[3Tu԰k$"xS3w5֙G_gҼt.%ֲ 뎄aaC/zՐxD<z=(( H`넖ow O.wP{ڨo]E3dϳ-س,y=h G)*RuGv82atЌڹT o\ n6cP 1h따\&kB Dkمa=a!J/ǚ\7>r-u6jQ-@S 6*i@e采uV#hZDA+8e`<g 'K2M p6׹,m2>_%C^_D{fr=y_9ru! yݫv~z1sw O3~eCO>!gk~ &w2'pC1#F3@IA'-N"< B||F@ WU}sRަ.c$ͯ/'.!/iÑr#oS?_v:9lG M] PmY֊,}\,lT葠zP{ '"QAG`uy9gt}NpIF'\a葳 cѣ+>QžB󿻸v#)7Xs]ufrh(g!W_Q/B0> .9+I<9Fs Ŕ' .D$B5hʼJ;&4@Js &wGzoKмxPQ+D ${(v (oղӄz;$uc ^͍yƴ0lDs`/cs͢M;TO84BDN4S0ER0Ecr({ɃJyE%鵍Bf( d4&7  ٦o#^4aWַ%Ћ/;8ۛuI Z$1G⯿$|Q /rӻ /vFu -HbDڳ $4)aO)e2'P'}^rvɘfcÐ%KڄVq|?qifJgJ]i8K.gty.U;QDe=*=?g)2N܆[I嵎3WN]˔g9Aػfo Z&9oW Q]ƨ(Vz6{]WAUЩ֒j!"ډ+h&TRAY`)NԖFim*4%UL2@kfL&L4Ϲ3 #F|@۱s];0j[ -HwehKBԅ# 9/4w^k}16 ay(bY$+p^ 4ז{8O!9r'6 _|U!nuFeҴR5ޅ$! bHa 16Tai!PtRy DcQC^̅x~HR5sKyO<`Zi.lBd(x QTaj/jl^I_>yoLhԔHr}k.9I*act6HI>Nuu32;p2DhT]`N2mzBO DCv/aP^ #f_1MsU䈤j2) rvf5OrDṘn"Ė7FލXl.ʽs-渹Y"1gksAzCd3brr& ,{]Mcz0 hLl8tu F]2:) EIMJn Mx 4*K[t–0$ARr" )p!%UrBRn=:Ab8CЊ|sT9y Q'ex ($QahTdFd_d[TQ1#!R&R5P#cޡ8.hD[6ːX$O1xq)x1Gg?DI&Y4ȝ[R (w 3%4g]ۛ힨@E._i~b` ODs&x5DN4S4Ǔ ZxHC&u?ِ)wҲX?{OƑz;CE/X Ĉ̋Zf, I9N=MͥjVoEVg:ˆI; RC~$ԇ/9!+ey uS#Qߔ V`܋ xNՐ=3ݔdOkJ|vtLGKkӁqPq(ϠXi|`[kL@*H{QŨPy!#+o\c|#KT{g'|],R_Nwؓ=Xԑ=B˞08l\Փ;x'H`vqR)wQKV(h EH;c_B08pݣD Oc jjA,HӠ{2w4 |r(4eDB˭K#T(sG=эXDJmp#ҿW^i)jS.^Fdy bB 0oyMW -v ~qeBWۿyкۀ+].@F̴6wBfW(WH^L.7os_Qt4+༢QxOz(fJs j*8Y`*-^oUުyCvX%RfIxfW#0ZSkm7Y&}%{meMꍩ\ U]i5s<gbkM./LٚP@͞Ft:D0fjFf7\/9q_־}zͻS=ƺѫ櫯=Y4^ln[il 0C?]#t_nn{r{. |7ӿz?v_7n/gsDrV%V I3ѣe I#S1絽]bD|Hpb0k 3$i[nG ֨;Y$9 Mߢ,ϷjH%xQƙIlTAװ6}3]^z2z{u{9W`|QN-<-5ǫYU|=u{`eLΤ&R%?9 G4GfmsX^ŭ:gQzM#ǩkUb>v>gB|UΝ?x@CX4~~DP*]mҁ`M^#}bmr(R$H"X# X|c "|Y*5pQˊ]qͷ8h\-#:od^P 8U 4hi8ӧcJnC%MhMt<&C$-jHA=|<=D2L:;BraQb\g Sx̳>p zxpIa0i VkA/,Xmտpx Euwpi a(8ڷsdnc^7 \~ݝ[ŪYVCT>9`XkBV@haCk >#pIR=W穔8hǂ@ 7WDԢ׫C3_` y+j{.['+W.e7co.V߀vzur9Y^Mҧr0( tv T0 2 j#!VH̍Jtfdfd&Va,%F) I<]f&M#lY@v ԑtGC)UdN/Ef3%tX"vF^昨ҵXk]n\ZOaզ"EJlk/]ozpvwr9]vLI&5խ&=DZgX`D7z/~+rv7x Нh B|ϩp*j/#dJ+g'TK.'(diu8DpbJcx։u"ސGJ4gfEWDǑX U8ʝLvЛr 1xiU{}s8+Mdw1PN+9fleH5u[Rgƶ#M{(&aՐIcoQTw4jSHwiL&k.2[N4( _Q4s Gy/ir'J\8f3t;.&.0%cELm-o)ժb6-h{TiaX.:^=ay=ݲ,Ğunp,Rڝա+){r®&wNkΕML٘!Q DqG8Ў+2&FÒCi{sƹqOu3DI.B8'C[_.l}4Ig«eA3BM=pW>E eyhJE@iFÚ=a4 ;Bc.utKII(IfJ[ )axO7pqcXb͎bIW^n]p[zPs0ńY~($Kd&SŎg&8K\1h&F eH %FO-Y>D*fH_ӫ-TΏmiDrApc%D2fSǔ3M,Kr-?Qk? ax~vz d(>M@ѥ<^$\H 0B;it#/pBLiN Dx"T!r6rj+ǸT4q:)gY.ưbqLNτԧvxVZTt0Od@(Am)Vhp} ޢV:I~@J۟`P`A0/!ⅇmf~X>F'win@fWg9zb$4 ys~IeנW?K^߼:!੏oN(.6w^x!t)U\}=Mg<܃`bHsGׯN@(r.d'D?dm!TpYKKS~Kat]}tm@,1"h43戴[#krd"4R!d8c3c =3hBDD vBfYh*beֿxĐd8›xX_=d]Y Vc+&9Bڵ&UogoƄw*ͮD/ ">^u]?_mu%Ѹb;[򎒽{u 孑jx}jӐha\z}xHxi9jH BxG[#IV2?sUò!K0f3}<ӗp}D5Lru5R=@"0z ?:% i^w+(|wL_S60qkv./=Ί+0>Ǩx?r έ5[8[xWr ;PɹOw`<{4Q쎣GZ-\4P:ߗc',Ag^»`MVqHFV@xD*&BiF{kKrS85=$^,>/(UӜC# Bez ybH@%m%c%v%.r}K45atϿ U0v2ZcGYnR, M9P2 %,]!CV0)\Av[moZ=j9 2q!1HMaGv7RCNRi1if  $ˌ4A$H:`g2$%YfnJ]kɒ+?6tF}Yݑ١}}-,hHh*}ms"##3##Ke3*Q^ B[:_6–KAw(l6_ S([G ^cie2j|=e)wnJ B./-R)EHEe,pAʺ l?ݔpsR.QM曭g9hx,77+raXlkI<2ke0zRr1L>݇oч6LYiMrݲObpz os|_٘֜rEt~\ݢ2\Sog]g$qUEaZ؇h9p$-Gy"EXJ06`P"jD1^P9MWq2\+u(I E(Kߍ'^\8kTUޠa4_"N3S0r5~0O7~8F101eэ2CMpY-SRD.`Ox$kT dZu> Ӱyod:ʹ?˳bG"geCq{(qL9t#i422NH?djϕ1vo1 3f RrDɥ@i'M\zz#בַ&}8b*d/uݡ2v,͈l n^ xbdyr0{_32G1iIŒ`Wa~ MM7,hn,Z*ժ-VCQ*j0܊j4ȶM{Ej-6W]8/jnmDZ4R"ϗV6@$bÁ}1' Rhy!Ófub*A[TnMs4&5RmM_>aBA5} 3"meU)6YF$VʌTL[bAG8^gK*A (kpM͍fSPБ;Ds!!n[+ޡzkZDM7s"%>_ݘ9Y'Du  NR 7Y1AUa Jf*WFcXn$Ce,'}#3l&y?bP%7sF9Ppaޛ1.zer/ZaPd5p0$͹[3<@׋φY Cfmq%r5iPUr1%[;kI`<'I>ARl%\&k+V|(D"/0H*p ǐTa*9wi5ëc:yQSL\=:tCOiȳѡ|spq)=Q,x##ݰ}Q6, ’YG6sa3^8"L xgpp`%YVB1k5e-uei("B{G52Ju,Gq jY: TkV@`TS9<Ì#ɜjCol5B8Cy#̒5Dnk|BKvc][c]%J8CުC^umْ](ZnnqV{R({ӬboҔ8. ;;*]OZ<1b=r׫FSA^2CL0tski0Bͥ!^Lo3=y 3[o/L|iL'=t7-I{ݒ^r7轂o^w 侺-ǡՌ:/G]Kꂓz9Ӷj…y^$,eIbrx*2#8IeX^ $U學W'9V|ԍG_d[19_Hufmm&po˽ 1LQ?3fLfưLs eZh2J:9/6 yB[>qJnս*t`tյjFqRRktjnX3nRRy@~ˆ4e# g"jb Q A3ǷYQjmw0KAl 6smWVkۺr2<Tmu ·JjYl:u*$ҥ=P0B5$~ C} /5Cic)Vx5ZRFpi.wK!_\Vl,5as#0)a<ޅW ܳYNSUor @<ĺ?^zbJI8L,cȁUS 2J)aR_}mDW]"{-6Smާh8rjb9 çO1ºb%6;E {P[9¿e! U 5@R$]$[,#:Q0cPdz!=.,/ &J,p-yuD`p#n*JH} :GͲeco^Tc.~0 ci8'@n5#81ܫ`wW `^F3=D4L6!,_&WK/z.w2ל7osy91.MV݂Ng_wq Z)tT.৭{@K$Ϧۼ:0n:ޙu.Kqfav q}Dk?48mNߵdޫ蒁,9(h)V\ |c~hlݸd8u&3*Q!R2H l?wH=O :Hܻ-pD*+D5K׽ ָCºH<)X|)215ɥpU+'\8)'yT &Nrii(Ʉ1&[NP1!bZb_i Ȩ+yCS#tbDd +jN ,G2@x݈|X_* *y⹨2 ?2a~"0 Mұ@aL6$TNN%M4eufݡ&6*8 W$ CnPhHGj*!fD(j2?t`'4O|: HR jHFJb&>2D%dR qe)Bui@lN]L|p{ypבgӺHG\ "ByDbq&}wT }U9ݹ+=zagUfykY_ǖ}M} Xk7'hj9:ege54L4,-#aUb%)L{^X BU") wApPbV 9G ~)nhг3PJpr Dk7^LkUJDB樁P!C(3 &x flۏL EԢV61!(ĈHGQhaAؘژ 1/D;gsU+fk5d|Ep p1Νvt;50ۤ"KwCEhўKQUHvRD?v}f?&KөZIgOË2&8)/>ڂ)`?yYMf tYDP?k NXp 0g0.x^sA.a>s1̝aBR}F;Ips0 '^a^ڿL.__sHUظk ѭ ZgJ)~Lc5WڙEwkfo_'wiϓTye;_2̨_n?O&vn%=zT7?}iFKx5&`fAАT  B5H`>ѹf.}J l0nW (,rm6_@+[[c} Wه?>hop w~tZ&BG8]r9͐|>̢`{Kz zCh$r!ho^ulLE/Ucr~7cjDO3OOj8zguir2rjo~Gܰ6ugҨ1.Z8?l{nNNBPW^GZ_*-e5^زpEjm'9B^iԺU݊*9:qٛ6EOQnM6&V$Q[LY(dI@ZX`קL>h J1 *!1V$"Vsle(#l9"5ZH"tЈ1wj耱ު!%P T{A]r^;:\<%׽TiAk >DIN4%[ ҴU"pCiI:[ C%qY-{څ(saHoߧq Mǃi:%ؙaw7s{$Fv%w^}KY"疑!tvb6b4/Dh^i`y8r<~raDF;ECHv:e_&6v|9.(P~+%_OX[LrQ$G@1DT-Ipj:6f/6:ֽFqё-ooϷ`LnT]_+^jD}.d.?ys(qld$>4?.f ceu_5L} ^f%'5ϴ~+S 0t[%mr9Ersnjn6~usr`d0Hr2j޴0*o+f-Zk:tf{; P6 qEY\Mǚ}̲TI{=f:>v:f #[q$qn0ܘtU7Xy)h|SȗyCs`ߨ[;;յaH"Aeyj>|h(x]_axxP')~Y^v\|KΡŽЦKw 2'w(P$[֑ˑ%1NYZs|Y^5ڙ&;} `1;Nl~m9eexI)_/glbBeuT`EkGFn&F2Q~_3%"؇ER}.Ǩ4)/2ȂNQ(yD-BK5̮`Í)\zqJ+.%ky tC`[6*Lpx,Eh1l5'c?8O|Q~,p<y Ք2Ǔ8eS@ *P݋φ|};g5g"\!8B%7nø ps88.ho)>qα޹>ևT}8lr$T5>BrX:AII)T"NQov?8)Afެ| Ƭ# ldZfS}t]e%7lc%"b @.߫B:Q4UVʰl`R~.voNws[Šm}G-Q탽rLr͑qᖣ H改V[.Ĥ.KH.2Җ$ xB AhmOށGMvf4~r8'LtUr igK%Tha7-ᣐZv̡,hz6.ݙ䤤Jsaa*& {]kI'Z:T `;!i>e>)%J.1``e&g~Lcpߟ.j2nh^pƔB~QޞdZpD:Hy>p=R>ԇ\`&s¥Gk(ljG P+Gw7;VkD] ?۪SA%Hk K*V= ehUŒ%0>q mj7?t-o_a؄ "?&¥wDZg B-x"j&t(.!KpZz ,Q&D&4txh#j]!0 &CFBc!!wrHu1:ƆsҚ?W^䪒1 Y'HB"\%"Z:I*5ݟV3nyepo.UGA <hJZ a~M> /H;4`h4l  $&`~XA=;Jvh# )̏ulS]`9-O>EQ# 4=+5D<ʹ|wo~-+|`rA hbĺ'-l GLx[Uo##sX}Waγu$5'XbFMPvJ{>&hXhf`6~a כSڟ8tXrsS~Pa$&JˀZG`kά?ps΂~X?pSfY%ZuKng3"V2sh'N׭Uliu5m]kA7Lg},UIzRϕ {/_ٲ*7Sr)C%:RَoUrlg`"X,G|,fq3gwz UtZX-\Y~кO`tQ;?Ƈ%m\GxG/&F WK=S }}s?$9kU^M-mJ3ItpiZr{O~!Cabhd ˍ<7qf9wx^Zi6V^77/Kٔo~|8a,|q~ycdL mC$XBcA X&K"{嶑$혁Tz;vLwKERllGo@ v;yUV^I< Se !e2EşTd,F!YQM]$W PpcFd.>q\]h {f⮔ĘF{`0}veGGrr?oS1,SB;&vk$ā^ڦy6;m EVJ!nԔ3OwZ\22*VTHڳ595=(Yq1Xw?qs9NH(NfI1D1ۨHZ@5} ~N&GOĝeMBPRBtil؂K[oY'l+>b!ԣic8Ʃso|M' ]›N*b=*>Q;4ÿRyvl~Ay)4V2:7Ԍ01s.IaмŠ2\yȋkyx;t38 `^YXPYdV NZrMYZ :qgz>} y:x%,Ґm7^FO"˟BJgX<=X^,ХLG;FZ1O%ĤQI9lgxRDF9Kܹ?)Ɖ$v8_̾X-Tr2CwkIO!HrDDVl2d_]ubmĜIŋ!).F5Rh&D"mU{<Qe?yai=ȇtl͑W0Lbns3)f+ ZeT>zB/CHHN"VV`9$ W$ -rʄ&{҈ߜ KGdqPSe>NdXv6KZ a卋kQQvhdu:찬k;Y٫)6J爤(Ke). Eθqbi^O6&?_]"yx9I?FIAgD.YJspͩH &E k~:5<_i^zt⮙9r}6h=0P+FZosh^NbYld!uw.?]{=&e"|HCI$9۷# '}E˥4^N޸M? Ej]d!$WZOǧ)G(B@CYO,F!ce l X؍ȗm[:wV"y\k#Nvꧧr| U_""  VvJTZ|aF 5Jlص`1GSgE~gݢ9-1m#"{CbM9?cVS%P"Ej>7c'HS l$ё6Hfa룦Kv&Ŏx] 4mC/%ϟsdiT #5{kG9b X y~zӌ;PYgџp,g1w5Eʨ ΢gW*M"Ů v m\SN[k=AǗ6+CZN'Jd.l$m+fLc^6d;1Ftlq`~g Sy2`䎂vPI<g ^Ah0x#`XyNLvйfU pkv= U ՠw_Ѫz>}\z@l >ɇc##XI:dϓ}ՙD֟a4qa)%#uȗHLѨ떘Kh3-Q|13_$9%o;_/)]릲Ld(< W-l<.~nݞ|ʨhLo h1ŨS~OQd`dF(v[Gᴧÿe L½/Qt|ud;`4u -G .(gs˧ljsZgbη[N /f(p-)o_cO/""A4Y/ >4lXU 'B?W|}'văp %{\c`|*%FK0,|&gJ tS@cR SHZ$-L "+x"dvBp* R*Mʰ¤ k~/5BI)$ FKσ'axG+2&mcb<]|+H\ mHBd&ZD#bX;&ܗLZq4>e 1gO2=Ff0lȼ/;"*{w,6e֯>T>XP;n1YJ.D~/v;wq+tQIm(ZA"NxđX 4􆹶wo:b:nw;@k'߻ʐ\:MU+bh& 2BS*H1&1B̵&-scҢ@" Wi.`f }^]έQZ` .M& D<ӂaf*Әq#F)ISe9R k&j$g VkFNXvDyiP!1TҌg4KxUVh]eM xIȩ8j:oW?T!UR&=]#IC1Nʙ2$? B\406ՠ:WfǛei>1Z/`Fv nM͸vѼA{/|x;WfPx]ORb34vV8Z[ŴXAp::NGaK41,YBUg`J7H8}u9C95c*_"uC|(n<"O+% 6oT5>~xEV&B #}<kcN qԍ3M^| >)(#BV̆7b[Ա6iqzfak^"%&T& L4g*$\IysZBsR>R} ƴGV;϶<:Gpt4,P_y7 vD"썄o/1Ҙ!7">2S]Z8+RNRaiXe(EN%Yʄif$"R$͜Їzґ^bӚ$ԋ@D LJӂT̵D K,XZ`ixoh[Pw%JZ097yo,F4]菊@I[5]{:1Rk5 w}V}nO5.6nϪ>G%b1湛Ҥ^KJVQR!J5fJ^5MKRyer1t杦 ZZ u! !Zcz|nׯzxhQ~mPy+jtsn;j,!|Gv2ATsUڊD/+~7]>HC2VvYV !z#vAeHb֟nJȶë{t;aF}ƑUx_ٹNKNMSma"LhPYiY;bR|x;רHWf8՝Y[vTP"GC͂Eqd`iD Z1Ĥvk'P,vWEpCuSjEԣ#E X`JyQ6, N'𖒍w[n._E",!<!˱ǔR  L"O-,h8X%yz7vR~Ģ,|,\\˶"%LwFMbU$N*83HHĎr)DKX@! H^*<쪄 PR<D*d+Vf*ކp4y%}sXP' { $N-|0`ey3c &3,a|#.rC4㤝P>!+?0*s.Bu6w73`.֋ dnт5ky ~k1LP9ByDVyE -i )EA:溒S+R h!*3 pg1K%qO= uS "E'><S_:+%D, 0G(kvnt(ZhBT>ouv*}uѐ/nySS9qL]h-MK&v'3F]_1 IwCw?9H ݭJT?X Qbu JC.P4bn%YF R R> )+'qzi{8n8n^*&h~to !QR<$wƔGep7ľ_|gguIjd $pҙlhv_.f)N7S ܵS"}%S ~:zu?߮B#1V'^0_~>ILY$fmcf-oӿo<Lm.O^DŽKoN.]iJ9LJ@9__;UH*7ްnQXRNup)A7VhukCCT%};M&*ӆ e0Qw,*0":JZАg%:%qCV|%9X%jQM3 U#{LđGaA bq,lr1%Piq;hZs WK4w؝@e=@*F7xŵW=IA1R ¡BĨ.Z ^9VrCZb_ 1j[?N` y*ZSشnrB2u+AꔎQǺ ?Jͺ3kАg:#{MqL GuJǨc݆EX7VhukCC O&7Zs4&G\*NE QT̒j( 3!btycfͩV,=zYO5!|$2ȚcZI?)㼿vGJ҂H"fR5ֳ nwLR蓓rPtT'D:QL@ȂLHBKiKI]N˷ !rѹ<>5p/fnF۷6xR'jU5Ň,{$QJ럿#^|7 O D]3)[7o`Q),!_^-Usj拯rsW|"pm *'gΘx!ӊc&m;Ÿ坊hEF NJӍq3_Hi0ɓŹ]{ҦT#:}C~Kˏfľ\2Y}^.a^+)[f>1C缡b|^؇yCc%h|7<m ?|ʙ!249^?.AdFtĂrC:>3'Y. hE}q$mBmL7_L5@?'bq2DzS%{|V aɛ~x3|8?[6TMHdjղCH$H*wrZfGf+BPzgb ,VfJDB2Sn<ۙ"JWLlhhXZ) J+3Ex^-͔~o߫9XpfEj`Ws@[u"(,ÅB\c 8D:_ku0>*$ I!PW>*$&Ҏ2 _̾TmA%#'/fp_RK7/+/fN-4+zm,:COn+LGCVjle o.&sst3Vٓ/&߄U62eb&<<@9ʶrip ^*ϓ7AJ~~FAt9H3-]bC\i]͕F? S9x‹.'+KB 8HR9'{MIL"LFXTFϟN 6 cQ9>&$$ۈguBg$:{!h:+LnDSƝ`c%b u0;1Fڲn?nP*u6 ɂfSxYTK)s 'x pB@.lĘr1 l agh@q+t D$UGQGf=zaSWҦ07@Ik0.j!Bn/K$.^O޵B%$Cb@g5MR‹S=!E=j-j8]Uuݺz4X8cGa/ .8HAJ.RrqPM.^I'DڠEVRYGҚD\#(59nq#4JA_ާte׹ʰ^ӁVӭ in1lvƃoH̔f >'NLa};-;HZv iAUV :4e!p+3rPtFtbINa^o.cws5Qs:(yЕf,?A{W d*NIl芌u :j"Y  IzHFywzpN|BC@Am%) ǒnXh~mGy~ʄQjM뫑$2jc|0@<E%J %:`-@%R<kAnbl R$ڶu odyxu(M!n _űͷ#ps5؄~M)4"hOZ$3 MF+׆z7[%ݓtOk+!1[y[vDC0t&Q.T~49;xތRζ_uX-NϨU;s’sVj%!鷔hڈN!@|i>j)$H>BI4u~( Ѵ_7IvY%C@nhD 4s TXu:‹/b3_.Pg z-rX,3gfϡEBsvu},ϊ >B3*=9xe\z9E]){ z !lg`DXU/ TVjDdɩR 0Fzl81JTYF i o0WɯFn:Pj䦃,=(? T!Ǟ1"t0SI2Nx0GoPiqtʟ>7Fk  M`Nv7,yQFBrJ#AsqRR }C z6UM0"`G,Fp7FP v ִdic`Q=eÂ"\3jf3BNEQi)i;F,m0bh:F L(U?|c`; #V pDGFl'6XvXc46°+1u)0URqpAx\i-QR91DalBU:j̲H뫱oe5*N34Km_zCQMo;2 PT~ ř>՜AP(-fpCm)h 5E*"6 Uٻ=BW;)dtj,UNK"  AEX WMl"Ci5ˀΣ %," ^ +47Tɠ# MPĈk2\+.1 E1(/]iA(xt͕&xYHAvȰePaTe:v__W~w͗=Xn( ,|o{BO_'c͕ox; ~ zQ ZT 3Ƿ!p|;?嘀r/>XtM^Gf6J~x4 Pq SƄ+Eox J:"UHבMb!];NL$.w\J㍎nZ&vOVJcwOPcS}"'J<tE#~r,Ӭ#x\pJZjn)xwn{ <)xRbZ&ac w* r VH )j`]s`H@%S%Zp_@`G6 PkA#p p i3 4oCA2<3T6"PO($&b(&MC@Pl|e,ڴ4S(,]з/$t|p Ѐ7a)Sˍ ;,"BP.GJu푃^nCLV L_"H|.dIiiX>@)a+L H;f^pLNIJK 3okǍu4*VnTGZ//?~99ݏ=I!GBfoZSUct@Mr'>#RA/GWM}PXT2lao.o8. 7qY̥~ж?i* =y*Kz`f8ZLCo.Wd{\|o\^b"or"xC3mo2&7t~1KT U1/WM@{x1JuŲpz%Ỻ`Ll)֔i{Ĭ'H*Q0LedȩhB瞓JbW~fpbřc /畻^6+(MF (g[& %3iWJU-˯GqeGss+r<p*w;txZ[33k8tC:BNQ(U%L l2)׃a[2Ysi/?BKeNM‘<Ӳ0TDP2UJG{7n:,&mZР" LGd4,%q4` >1t)rʙ^Qt0S]M6;jT4afG-RO{Ӻr:qo%^SX Ƿe3<*W*))As"zk74bf0?_#)g|r8oz@ wXG!^Y,TBCgG,s[1}]@>.faL8W_|JJ:w1oRܲ19C rgW*>I8g5Ncx4*W.FbV6Zݫ g;Cvr_u !98ocpn?Bp44[ m+w4aW}xl LM>wT[t9-9Z!/{~M:v}-3;Pko f'q5G0;9st;Oq9uyp~_ӭ4 ئkdoIygZTr\O8cGg_cyw2љV&5G;ѣ3p O14뜑mU4y|fO;bt[>,8I-296u0;'>G4C#]u Ǿݧ 8)E(N)898GڹIp=NQ͍ωvCj’!5~XUeC1}HDnV)) ĘuijPb~TeS LUY[eSoJe3s@²z:L}| hp5dG{~D `Dwv]y>Gޜ#:y; &_&ۏ*X=Kae]]\~2^LXdkߠcṭz빦sM5AƢrU빦v5sM<פZW: kTUO)I|Q=7rgP)dَ=;3 iwz* Es—Q$u$K өU\r z m ڕR{ew'w5]M-bo<U#)y[O4:noBeB~d5QK_i#:5!T[ d!/\5ATR}?;qx &Lxl QK<`ɕu"uyT1!"Z9 #p4e5'i#CP/dR)a]n]׋7(+a맮JrGf<8NͥY'M w TD3-ԾߤVJ.^HdL"|f:vQM& pxK|;O?jf9^CՓ*(g[rg0G{2WdxbA0bܳiw"[3-u?m/N'7M+&ORunn:oDtOXD;?di|?[(])$tп߯ozo{zgpd^|oZS^nnnw `S"PW0iDL Ң;1NbCr #g7v(_zWƣVSw|E+o *y!0ԙm S%a67t||||Y[r.AE%BG!u GcHwmifQH̾@nn:M5u+ w}߽CU4sԛޟ "C3,Fwd?~A&$[dMjI4ez!8QJMM4k72ˉKu*JQa&`#1 cH+lБZ76XBwYBwt2<{Qڼ!rD3A}YimA&T h=6N=f=!1}C}}jzdJj\Lz8VN(XTmKuԩ ILpX.| uϥR˙*->z5R]2̗7/i6ܓƳk49*؜O!w]sPw6Q\`LԎbÊ#Xd,H*qŀQp`6 I+.y`2ۯ8ƑuV}P>nͨk8qMrTSP]_)&8mNLBJOZ]w3t=@ltC! .4%[ b{wV[ wȮF)%iG ]W&,It1OnI&ٻFrkWHܗ!w'`d<$kfܲa$%Y K,Vl=r}Yx ;#]a'F)k2%,sH#D3_iri"(ʞ]$:M ~St Ug3עɀ p׮/ 7s1 8G| D;QIh{DpW@vHiEfIbTa, ~"jl бQ!@DT6$HB !l,L ƀvtis՛9S ǃ|z&W I&h/Fv3F ۨ歝n q4z ӺDRkT"A;{v9CSS:K&}kdw %IZWN30Ӫ"J["Uz|Ք7lW BlttP8A2–"ҙlgJ[ɁBRrvc$2/`HQkhNէ僎\ȱ @oT4m?7k7 89ĩM!s= nCo\zҐ$m;ibb%Q8'%59ƹ|eqA%SО&vo H14\[c$ !k([K퀜o4LR՘;A0(J -A[]2֝E٭ +6,h597 Ω70(%5vX)Ztk80?xAP2-F1*YJ(=x66NHY+ DZӱ_GrkUX'qJgz Ytwu2/,<㻩gqwvfoFvQwAl 04Ɛq`c +ZROךEi%ELjƶԐ@b1XHL0Wd Q h-02TMt挛ci4(f.Z8Y5WvژjA9 r\_u\u5Iuå\^:WudL)΃L=ӠQ]ꌂEU~8G'7'2: _LڱK,dȝ)_TLkfOB[ɻkhE/dC@p_d-A@Qo H%Xo Œ; % GPAcmh/=0Dn-30 sM߲c˶O_q'_e'_Yf6xAUkm*T: X - jILQ*Z*?ODOMnȼʓm\r]ڭPv;焠ZY܁ɢ>hh8-fFF]t[苀}7XVڙj=jDp~xY<'c!%C#VYͅE/{dkscq홛f@K3-tVhXskM\ʏz 3} ځ02#e,Vj7->87qq?Ɇfσ^OY-d CVBFY u}ٓblM 0 /KH5!^ڄ {a5myw ;Z'xihE9g_gsV|22wӨ[񭃆R2^>hkqBv6=据%=[Z=NPƄQ^TجnA^ũ͇}zh`fNE϶A!(?NP&1> ^}X\f:_n~mpU>;hbHabV?{ ֓i)4"lgo2[f7/4qE` 7:}eа3Q$lA%+ʐ3wn|C:cAn{!7 c [B@U \y%FZ9֎.jyq +GFAH=INd;pS';M ưcEM^]ԓv( x3%޾)A4$j$C7=d إeuKR(f:0Os'56S[%a v[Q[ 2^NE2ـCz~œ`0qx_6iqٛS}d'oOuN187І[(op_e!DFƁN6뜼_08-27^hn Ji<줮HվaQ=^دG=}C%']CC>N"蜹/B\-iOE=u- 9=$y<,"dmDr? 9-2kvL9_ ^"Sx?'BVsZȩ(p-lٶsNjL m}h[ԜS?ɰHPb)kTc=yH@e?7bAR: W͉"d'y _85iy?FeHfꊎ_GڝOj>⁐Gߪr :d>|uoh>Ggq]v;P>P<#HNAT ) eL5Np%pY6pGd FBFMpg S]Ov20 ~&aGr?f\g#(¢|}4돣ѫKHkNY ުY*0:` Uϫbtp/ni6FQfpX1TzX!-JRObŬ~ ]ƺB>kj`#?|onnGoV}mըsX^|E /pRx%BH;5W'Ebo9$5&F#N;&13JqCF_I-4,&?-uWkew&I_сf#JeC) ac@i1Sa9=@1h-L@ jV9,WQnWѻE2k Zөl vE:5_8%[%̺yBZD0Lc3bQ+`C!_;$?~Y;Y ?qo{fۋ,:{FCҠ鎱XcEcEÛc/RgQt~sT)oG7w eN#GԦiu{iV^}ѓPJFȨVp`]o9W}m7A[`L&_faFzd'$-YխQ&YՃbhD!k͈D1'jxTL&KM5 DoB2YYs7/:*ڢb9JG4o@ M&qDC3@%kġJ!dᡅQΔrӒY婴H*#"#! !TC(BRL)j91V(e0.FŁKNΝA!|#F%B^X$i@`L UId\5 AxBQB1\q-L'"X-RsQx;Y9aelCNStY}UDxY NFhid8cgY1oq2L_^?̖t5'TB;uPPI[`fowvuiJM#ZźcA dм;FlAOU>w%1Kut։s|DNތi2؏b'YR)vp>I>WOg \gxIdTY3D4n 08E/-9?|bqioq״NJak|>1?  "z (+6,1?"i(0o$$(%lT^׽Қ̗J ǽ8gA8N{_Q1>C݄kߊ5+3U⟷Ad^Ɉ,Z-S3'}\R~7vwIe#9^L~2[qqܓ.4LLC}-y_9_~ilU$5~e^T"h5//VAȭ4$J8OFbB[ l5N;Bp! {mL1~C"gڼL7 % RUK}mL <Ȼcb6ySu|kU^X >t}m3Ko>^X ^)RuusާʂnĤ=??FOd8VF|4B漏OXjn3~zb~&pmQ}wu6d@ \1iCi괹qgˆ{3(?f ('Ս|}3<T$n#87yB55|7yւIf| ӎ0q3;]~:)IbVP#_7}Ԛ*.Qqq#p-s*j-Oe#%-ާ|ߺB[)&Ѹ4MW]*t̃F5"Zpu6DH*qDY&QGCtAD&ٙbjTlI0ϱ.ʒ,&lEa|^dy,d$1N Y2h%e3)[pܸF 5Qa.\RrCSLʞR?Wo"2WCLv>YKd⪓iަyE9`Z+7d9C+HFʝSz lr$Z(UL@x#"H> "8 Et_鱺Rcܖ#^D>s` Պh)_ՔKZʽ8fUܧCai:ѭGflUU XU[ _O4tѦ ?G;y#;nGQy7ɲ7/++NKݠx}d zG[?7;㛇s}Jex ATY\Z ca߄Ѻ[DA^Ӽ[wNz~鿦wtYmf>;Gў.gWgg@Q?xgF;tЎz3|c c/?DILiVt[\]/M@Fu!l/w.x kvCklIA W(CN#_c -)N2 95[y p}Dgv:\2тV~ńrBs8,>#"G ({nVj*V}uEí]*.IZf d.d|W/\wVȸG5!{X~s;gq3!+Nn+t|4O Ng>vM LE ̀0^do=J Oh<&њ*f (#SIb@qfoҕ|QA!ȷB`eD!Zd9<'\XOCtqv{H\ɻ`r S?(3C9ɟl$<=UȾ&k*"/]M?Dtx9y}1YKI-frr׾ta?զt{=Rƞ ;|njAN[l wv ZꝈysͷ8O.yVH=c v, brs.:BTs3߄mJm[ MPj{~i/yWTt8aJ9<:ڔ[H g8)2 uv "]"R3-Jz*mCs@hFח0,>q]-Za<-|Qe ŀW}-_}RZBt>G cGQ!b|Bai.N:v4Tx mq{ ok.]'7{,J]$"KEȋ,e,&2ICyV^'M \5Ox\F<8MRM;jTi[QAK3jTي !Hi$ w$R(SV"$Р^+VxɄɧǂkA'GVoc(@I|yۓiҊAP2ȶXړ( D<s)Ü$DUB\j 1ٓiUа3JH(.h# H$A %BrGI [Jg1VzJhi/ZHP/Kk`U`tR@xYi5 wV/ 2-GOl:bN˟%J ԣ{?C@/L?RUpoAU-j㪖VZP@/UIDn+[9 wR1H&gvTeaZڽIBhCKq:W9id_QȾ[᳊gpIoTBMq5]rU3^@f4RdsN! $|XPu$ L:BxN7ReeB,x?Ulbs1e=|qyu^/=Pů,@va˵|[XOk7^'?"RuuT唆 }ulōh8P{WJjϫP沗*4S4j ; 16CJ#LxW%4Qj[q+Jb&SSQsJnWi N&"Z8j.U;{D:qXFSABr-J2椒1ZE()hBS 8Qm3Qyˈ@$0o4r>ĜG+hD} i@5Z{䷅變S.zb]_FZ[jT/@~U:`׆enۡa2!0g+]jلY.]OA@[JTdaAPzg1bq謞1_'H @vQLr잯Ns4}rOҗoh*AԬZG4״,W}7R5VM:^{8+(KyrC9n$s @#bfۈ߾Z:KH` .!.LLϲw\|=QY(SsM'UNypD*:NAu^oJ&^oYvMҚf_zw祠$F϶WuYK36z3v @2CT,(x 'pHh(݌BJ,P OZ3WG8D}%)h"Wpq[]Jve] CY\j=~w󧼆)&eͯM}/\*ұ4X&z Ro`bMfVd~9j/Z H@$)K~6u!SpbU q>SF~19M\Zu㇛+gm2Bcu8!+S.N p\E[U2osV)Y-IWW/24Vy,5 `;.Q4S!v5xN}?hNk^ ϟ,xzեn쨎d޲fA)Qg/~qN-pc>5$c'$cO*: & DρQ:F<{>YzqV.Ct;:FUPB$IFѵ=c0ֹxl7#SՊO|P#NӗavH An\-9-EgRE],E>ϱSM8!f|1Pk АB;p"u`"$pr~o&):qE vk2A1WoOD>49Qco"MN,[9TIBuܽg{4BקX:aINIVA_ec딎K;&J=X :KO1Z#IPB$(i$`t\峡=IQ.YF}&8G͏&ysjM`\3SX<|PE<ѝVPN(JTЪv/gqTh1Vk]_;ԍ}pVǍg!l>m5D=}=:TkڗϦ!=3!xwWgb2D> XΆlor:V&zݩy!ͽzzY?RD> )EcI ՌbtTz)D}Sqb|ʪjoσlu :[<T$f:28mʰ<86@x,kAI_> V4ƈ>juj2AKh}ZBsf.0,BQrU~ggߩIR,E7VH1F*.ВWx'Cv\;VY AT"\ow/r^}]Ľo#ܨ/Y_t^6ξK|=_]^UC?p{B_[tq5#{3W˖ُ- &1=LsrcTq UE]D;Q%Ҭ J˂Sh*^:9[w|yiE=kj\7EyƂ/ Z, X9},n۪Mh[}fqzT-@m .ƨ#DP%4"@ xR5׊oPs%c‡X)ǥfqvI v $0ǀGk瓂@)'^Xoo8k 1գ/vc wќDBIHF*dap)Ol0E C jS>4@#DXUټF@3CW&aŽitTńf,IZ\b8y.#r% '_M Ln"b4ǿ (./Gl "p:% jRp"9#aHu=U3W~e"%r`p?J-C120ΣZ)/LZY|mYp88ABFJ D%|) QIIX'ed\H};D oBhGZh , ))-tߍͽtK mi"3,.\@cA,U@=t.CD¬xS+`A׮>C@E 7T7鋃1] Y)P(\cj(${Ù!e({ABМn&Z$J29 %46552 (qkL?:S AV.w,(˂¦wDK -)rلdm"R!߄9pQ  Y9p]C"#B ȩ pi_DQp Y\F'L& AӂFҏyl41 x\E1-Ѐ)=vǘ\ܲL|RIY<{_mѡ!h9F<ۃW˒7[1~e}y η+=͠Мo4iQ{^M nB91ʩ4Ay vzk*b⻇Y>9Vdٚnf8թKu_i0/O{OM.X~/v?}r*K}J5d q^ F>5-08k|ogO-pa$<|9hmh$􆹶wo>w} @W2j+Z3 ]!B3əA}gn~gepeNp`-RN [ "P9敶Zax47p<1sH=B]l 6:j՚C-aGs\(F'g"\8 1$Ri|^S;f I20j)'=Xbֺ"Aod`QFL^]LtRVz5YA7ڏnR+[8(6}=r |sa[hMgF-ck٣GކNRR_+ܸw sIE YcHqE[T1s[M ßtk:v&.p2zu[gV<Գ8x[[as4nOEdҭBJ.8΢|G_jRX(і8űk9:"io n ADOI)$EӰ!Wjyl@{?=n\7"e;8mފd > UV"[ԊY߷ؒ75Υ[R@lI-,Bp!nQdٰ(bXǖ3JnD}>vad9cZR3mB(iiMQ(:\Nl:j1k2zAa;֧x{?lDA-/qCow!2%V֎*Js9ܫ*o\[q*e̐ݜبTIJ:;k(IVK:.:"X[2:Bt=}Cs%tZf :#r!4JWQ*_Q W=7r:/;Kv=koś/{0@+Mq! >O}.r仿G?ϟD{Cb[><$0b+S%<}dKZEdH jW@p1E=5_pj仒xC_ݖ]6g?.fX/}ɝm6#w4n/M,Ϛ_?}b ̀bej#ETƺM@_.{BI)(BS7}.DDrV $g f.Bjq:50k<2vКY1:cWK~Х[8~(G>ϒf`O9 z$N؎ 3Ƚ g;Z 9XZ =ף7b/7+13 $ccjyoN휌FeLRG,05JqRN䆻0t\,u&ŷD9 hhېz 6"EJ_.Ȏ@5<%\-$"JbB)VJoC#ruMi$Lc_{_uH #c`e)b%eSFk^R֐,$.rQ&4LĨerV-I ^])ڄS5XV$J!Q]]j^֮i-ZFyL;6N8Mg]kY15rlf\Sw ]da[}Wu#d(d~^@E[8d-HݕJ럼TwXK /tg0(JG>o5EN>^.]|v%ٛ(yݣj<5&9`>ƽփeG'?\6t>77a xyG)N7Miwd@7KUnT̎\*BY͇(dA9IBF$07,~ʯO"@q>\& *NM?zQOM ȭ+aE1V٧,4)BT}hL5*^d kw18Tʿ_ڃܵ/ j:gݣKs=j.&FOܵYQu y"l"#6\ARU 1KpuNS sn vB4B+IwwPrH!e币$x>Z!a12y%!$CJiL " cּ4_0lkƃKcR2~D,x6 B s aD[o YՇq1uvɊd{(T+/^܃䜍m4Iu!5QGoxqfhwíugՇ=VqT$ V#<W#}zrnctz"ґk'=w;>u"`.'dGM%f&$clr/~8./vw`'i11dj N0d([C>X~z+.(Jģ7v3_EeXG.{MD677ӧw_> Z_ wo fD j ktC>1}{N ,h AҥX}Y0k,l}``LKw_^$~kĄ 7o=d__VXˀ -((p2zeC]YF2ܕ2^f|KX9 }a07HWTz.W=.kiMVyjn1fUݷ)ķUv4T ӡ"tI*:R,4n B1fy݇Ts;^ bvep>G٫e]P|:Аd' rHE&v< q:ӑY)Fbx; cO9ldѽ[,LzƐֽ#]kag (-/kv: {a.zRvBZKP̪#)w{%U ;d9)8~ӷFkfHB)]/PRAʙR3%=='~LK6/g@eB.f(@vz׼5-fݥ3Sr _;M`4a6k>* )q.B 5Ձރ$f'QV] )I:(~9᪂} ;(AJu\0RK]WA1kåM#m7i9V/4KT3#隦=?T @,9:_-%J H4C'` qҙ"|&W n:kUO켸<[t&bL$lr`(y& ҡ+LWVm%]+.VC+.6HiE/*\XݬbԖ=^ زyXZꑞC<c%OC('xthjioGԝs+F~?ڥN<%LZc%^ FaԈJT,g!g p,$wg!yr&rsrڇ,s &'u#4|YpPE]u{S4//YH`>e xON,,}D))[ULp/ gV5cTS薾rvfZ`5 5Nh|(Ș/ *IDF4* DlVn\˃`3<U_|7شeW,x*m S*|AM.Y'nX+& W7 3-%v-m/~{Ϲq3 G,B+p.((ܚ7݅ p .pV5 a˯}W G}NG_\_6/4\~(r"ry@>l{E|F#K-|Q,KH6`Q{#,\ga|srKANnɏ:[Rxid16r>xr{ola,$Wh=GDP‘c%M7@"h !'̒R[WQ'fw#ZM?zP+kEJfCI#ͱ9cқ%ONhϙa0$uH4h뱰,EU%EKK$zmOFƦ3ZEũ<'z>I!"'%%WpHy`&^dW!&.l0p>%oM;G&<ЍC3 GG р"H"UjuMYbVH2K-DS_{eu 'N'9ݚ(~+" ZEFYSk˙Y oh#Tu/ n\M |_^Iot"P{%ʦ`fZbt=?τfZۜn]%O`Rc@Fn8\$/V6J8dÙ(m5s[Ț}vgp/^.^BJ'>lYPŜMu/-¡%>]5!~h-ꄑSOv)޲B) ޤ}mI8ڥN<)VIWZv ƀ6# x@G0a">0Vz~\AjSusd$,٭eK/ψ,g5ÆȲ f%GbYud$ezvEiB~iudY"Rp&--lɻiQ#Kܹ A1f+q3ɍ4jqzxbp~ɰ8#eςWjY"!#HYVc/ DB$,sOb#neQE*}G/"ivZάRt̚LbPPҥCdL1R& Χ`#52b%q1?2*1ZEBdҔ(X$/(#8tF)7DDj@+Vi7tT|nƕ$usv 7=]* .ղ 8WE$l edך0?9󳿺]?~7 +<;(̨@/חj\ <_OOUrqdI56n&My`e.j5m m~v>5>Y@F#0u3ͷ:m#cDI- rm@ MxAib J,oqi|, N4#Ct0͐D#9EhqF TP 4l dS_@A+$ɴD#$#M*˝J.K~بyx*6AtN>Nox߿7 Fо/HoV@9=&"zo-!P 0ae KY-L/߹[4T׭VTSD(t7oߝj&VNVsh'BsU& r.כ &ѽfwZjá%o|ܒ"ep"ț6LK;YJ7ᾳE ouܢLvxO CwXf1prz¾G b߱0ЍOg[%n\dyN6|loғͻ\IIB^Fɔoji7JJ%'{nN;h=ґæ[|@Vp8Ts]]{nN;h=zSE[n nuH ѝeJ4/pN.gDs XH$XnZ)U^; u>:M>6?}sߙͫft%gWy RgwL^TeK0bG]Tw5%iWkM8><jWnb;6۳ٟ([~ç;ڕ+`.dYI!k"n([?ո.W)SwW O?Y*,_q_}it.kk/ Erͧo{{׃&-rRTzzB<VC)l`I8d_,Yb`ISr ~8V lWL'_K l(*͎¼ -<9;J#h-i`jP7d+$)klddwb&&L/.?NWW7%"KH4NdZ]nL%b& ڸ'f5W-3qId{̖cGmRoݙ 6%T.imXiҵ I^yV5/BF,~n`0QNc3 X&4bV('-6!-[a=QA;ASC8mY+y3gR0$\+F~* \hR7ŰOu>Z* lyǓ5/OsTZ b':C< uT"UFWۿO]k^qO\ŀ_:t Y+VVmz m䃚EEԃ W"D+swHV 2!9ޟ q$I-Megȑjt9w wl djxpTlvWXRƮ. a+y'E*1m4 !}LwTlTʅBf3ڇbjGQ13 GԩO+ǬW%$pk Ih!E6]WXO|SRbD>; AM< [4x*-6"(L@3,42#LPǔ%3nRi /D' f8:da\/YX :QXXľӮ0:&e[f!X=,N7ƈs\3/:|iq)C)N$K(iku%eP BN-D0T5#e0C-a'Tqa؄b`?T2 va˹Y4EEbJ|WvSl7_XX^voG̱ArHf6+)c_ˋ'a%~{ gdƓ)3֔ᅦp7yO7}37Nzw?Myf\ "vb&Ms7 ʘB'k|Q(u;A@Q{Y!XkVi@ ~,xU*BLӸ__>y:9K}vtqMG0~х?gLLLL*3q:]2 V#%v\$sg:T3`,asKM9?yj}^udlP)؝jU]l%eXIa"8N?L3PT) )+A A?|@|.-t[}*& ,4Yb4PiŧL)MdP {H0R43ujReJu]RS enkn6i^;=OcOdtg c+{83.-m7dV͇h >?_ۚ7ҋX>C7?kȟ ?yZ1d)J/|c/sX_/? ^q?ds'W,XYk.Q2ӫ P.b1wTnD,JNTƐþ K@ QHbHTÌP1)q>o NR5 T, \*k90I‰%wA"x3*efZvz2 E频*6VF k4J<~|lZNA)GJ5S8nTKp]dКf 2;jF a_:P%Z.t rVg v&܁X!Qbapi4T+H3`љ4CtИNNƈZr?Ga[Nfr5nv%2{g )ł- 0F~B$d{Oȯ=VR fDkE?_اY&C򎃆\C yCA|v{Հ §5AVOU[޺SM멊[f2jQ&"bCTUǼr[Pꍊ Yv0D+Z+C13ZFDro]jE3۶)W@;nڸL7*lk!bC5Cp} 3 w Ze>nI0$bےQCaT6|Va@&GE"уC`(Z)!dHFG ~:S (Zti2k]y%U8W6V匮w-U&Ok"Onp8]巯>?ypd䥐˫#Vc ҽ*y|>qz[z//ځDXD㤐ԱpRHzK ٳ 1d(H,E;Yr^{ (}B̓]X{'O&G%2,=C2c|P̽G@{PB4F9t|$#,d$#$uR{Do{C=1~c\K[gw۲5 Pw{u>Z_~#V-Uea'cHIOpиO?u}-rn\gBH:6$y|jk|Hen1!C9;ե_}iB'`ׁZ8牡} ܇xԮ}Np4'W?k~4 'nv{DL݆4`@UN9;쀙p7ˉ]OǸוzƇRŷQWJBÇ0`0}=n-'w(7szg8R{k˻__5+:5˺cBih}*rjyVzZ5̇ %8VIBd@tREdj{ :7 Yhc ෡f5h@ Y*4+U &ITE+XŠ/Z7lrP ɸ'Wfzpd$ґVȒNξm0=xsF"eӒ0δkG\@YcBN/.p>DkD~m<]}2J ^M֯=[g?t~ן_V#à:er.FAvg{sQ7'Fɘ9[}/ )q;NF~#;Uݥ>?QK#orldud Э|cM\c[r< Ƀs2RB-xIvnn_ET#^N^p8JqJsϜjh=5iތbG~a@[g7xD0eqq-RC'MH:6BVk>H:HY K$0$vVkzYT,?f1$kWZ~IH<d J?'-5>;Q:b5fD/o(e}6׃Kt"H AH+L1 l 3"<Ipހ7֙e664wIZwPi+MKVv AFf%5vzk OHݼװX6vPV5{n~BX9WzK)a / n76.M97#O,=botMnjBiu@$j'8.#UR:ͣ>0V{?ca;bF##Clr${f,FjO*U[=zVmo}flA\t2"-ٽ_ d^f-FDĜ~h؏l\Q?,Q!xWX>!"PiNd"4 +%l†sCt ˱V9Fxv7IYm6z (_'WS-Wl,Q=wP qPr Fa`R' ZW?X`JKQ?+X+952G cz٩zLŗOׇA|j#$!2mn %jäϢF9zdրR6]S)=hS'K/ 3V6nw{W6v:[`qp]?ʋج;~fi[‘Jk4ۦh6ijM0l%t0sSo.  Mr4\( |%9:"lAK3Q˼r#0MCZTۤ—+-BPok7TL?UJ>E?ZQSre2$ e#d)%Imu>9鍲r ř[ըzq- H$aBmɶUEʽ'<U ֎ikֲ,ki?X;6 9+$}6-^?w$7cUDvnDw[a|{`~rR̤AC0[^uCaNqqzV[Ϧ Bz%A/6í⑚hV3™Zp>gNϖELG|yUOo޳{:3 ]w/^}`qyQ{R;U>xⷯϒzb9BN?ʯ>~|W~xTf!gկΟ>~`?'?_֢<??d{'A{ݔxey䦂UwoF=}w4:$E'NK0#;D%׺m|`;_~>g$nyhb!74XXLD61TYKxg8Z'ޝIں}իQt9x:g|r-8Z\۾3 cI4O@c Ht,ߨ˒XR}te;ܺW},$Ϫ5UkهKV *$gp.RJ9CQ;O. 0Dpdc?4cw6eB 'lojꠇ2ʐҮI8l 2`*(w?^<_*_c4hA?{ 14{m@챮tud<hX&GJq*6F¹ 1k=nzi% Q(W16.xHE<.߬Y<.ϵ v!Esr*`)i"8a(i@xA@.VP?s\qx 4n.r( ,z3oEq $U4}3+A1tF zԨ€0RM¯36#28)9Y( x~3uR#^,Q%2}{7r\SzT -:d(M†u^! TgtNfjWo`oǭ(< 7q`^،uKDzWNQjH. JKS6Cg?ӳo&cL2*KX8ɵ }hA TJx ;ٱc"FnD5w%7c|Ten~o"~鑵}j[/xܦ?n9bҟFҟFҟFҟINiƝŃ4U\o.X$3v`:%p3-JQkhQb"|7mUG^e6H<߇tu?se//*%;` UM 1 )݀7i֮mO:܅!rCE}i3Z% y&ޛ_ow]Iս{{m=I1?]oocC8ڱ'Wίժ;|UhXOađF$S kU6cGDpU{|6 pNlcw"{}/{"2=t?u|X͌3o:XKP.\\_5w{F-W&`z&?0$xf~aq"y&> rE%3ѿRSl]jE p& hZ(Z!K<Ņ_/?eO-^/`EqwSzp.xnyü 6O6~FCcw om!e#;uFRÑey DL2G!7!qJ%Zt q Cе/탱MU<u-Rp#b;Aӭ'*m: 1gwUV _%Cb̀1Z58P:gu{E`A\tO$jkZM˺</  T+0i{k+Co7gFㆽK`O1^3zB/SFgN q~髞YL4 ^ ЬaQ&g̬^4 j쵑yRNp5#^xC #p͘6'Q.q{u-:M\Sq˛qѲN]=pZAM2yn:I9[;ޢ6R~=87(?pJiL:!mIW-=^31 DZhRɈٱƣ=_5ejāB+E<# Bџ#oFЃev4%/ް">i(Q>O/6tGڢ֝zm6tq$8m;M/cm+$Vw sjH MgDI RJ*Rq;t3 )/Mv%d$SAhieKMQ3DƮ:n׻K@w%J|)GHDqEOͅRX3t(@cZa̲4[ki6qjjDLZzS7AINZocjAT&/zX hH1#fk6w|Ln #bdlt9 4 Z"|U@I%$V~ү\f2z&KmX&n8͓ӵ]_F|Qn6<%9<@5mR iO P]|_?F|L_J&(h( ?V& J<=ƻF3(w6U'SVoD?ZzMҸB1H<:⢹Age#)s0ɂcB(.I˸X2qn@F9pH?&nܳgVSx2R@rp1G!ju *g ɳͯh*eў-P2oصߌز W [8kgI^Ft"/ co&leFj:Sʬr2ȵܴƳIs k%tU@Y#NE4`J#9ZN?".]¸VHTz< U`H-U򍃗NkR/ ҳ(3J eȗ0dm$A`Wǁf*m|uF5([y+ Dɽl)x>ͯجy0엿<[" wЎYB55p1tkpY F8 s LD'}%j|hprC݉#'RLϷ]w_2 )cJ}2mjPQ\u9bVMƀM e.г]^a~iE;\"ׇe$0/@Fi1&#\͎5e1{)$){7Lf\ʺ =姻/Z"+;IG,lx(ն^ \?Lct$#=P+jq45^|V|_^Fe 8\z!6n5`@@>RGsąl 7z1ߣߣ ,eƅwpII @Ϫܝ]>M2CǓ:m\~9IoZgi/B$H=|  ]/10iq ^hj ᒠ)B4!a]͎5#|ޝ}1@mPz)Ԋ ?fEn*UǗ/Wqy8zx\Y- B !MueWwf}|[-e=|BZ'$UI*TYxpt:5>+z'mBci̶!ٱsG}=a[(1k+G c }L{mR.xa^V;@RG/ $/0 ~bvml"F/U f$#ܔ~ُ9m 6Z 57]^1_<?i<$Q)S)IsBr~Kg..F 1єSKcn_[ț5;*=apdV[sEn u#{~|X3XQiaд(ԇ]J  3uXb#G=`fg U0ݥsoI%F O<"r<<{DGIG+q@-bl^}M_>LJMj㐋Et[._z~b$873#qFy[Oq~~?A]-yHԦvК_[3}40yR\$_R5G gL9NqyeʄF:pkRF cIbg<}Mߕ:|=_qkN*Em"6l)ks覽{t34@Ls r8z/[/ g8ϢvSc$Z%A16~ .fDMWlq~5).dmú4@(o$B]cgN-p\)d1eVkW'nD䮲-]HeTrT q&AV:-7j )ǭviLƧ" )+jӚـ, hchXpтY&ٱƣ1zr=rZ7(8YwՏjJE 8C`XTN~Lެ1q+a {X\Y2-4,Y( Gm%GVެ񤕇^z=O% +c jL'z hzCeO_:3g*xG n٧,!'$s\ mOW傫UUlbuǒA%&X{䩐Z:+hT 5NRLJϵ=6*Q+D䚂i:w *:ێ?O $87 CXK%jp<-nq%j9Ay4MI6+$L2eb9ÿMKe Vst4MMBLk,1bƂ8`E^[&TB1^)YJ{Ƹ%w s.2sN.) ͫʞT]Jќ?'VAUc"w,1֔J )Ua[Uř6WEks&uU9+E&(#ÒoMC B){⦹&Αx񬚟!N=\@s%X۷ó?DVZN)'O5RCH?z]7z:IUZ)Xe%6N 1%^XSSQf]M4LP.YMeiXV0璭1~J{0&`H>*PP3^$ E^]FH)+f{(ϕUKoˌߐ$aBJQ3Jhfd30MW1bg 8ǹxQMU  Sݷc; ;ÌCMtg)nNǸr0^g!YUe\4vNb#NY$2 G ~s Y2-S+Ti>;onϐ`g>v!trE{@as1zc/7mo7b֓$Hvj!Y"5r*B!767.N\5 lРT`<3%}O3#Q(pN`"UXqhJgVyjk]MaZZn?B\yDŽH ` CẢITيϭJokMⅪk>YSB&9ɵ#_vBfyE,r2N+! >EKK#PfyN.5 f9E٬OGUg)po~W~r\:evau.w.%̥ͮ9b7ٵ,>B HH2u8 mיw" 5O*0VϧD XDj8 |xɗ-A,d=B%xdS  NGzfVBbGwajB܇qsw5z*D.)ẇQKߌʕk%lbP|陾{΁|i:e<[g' ?ۨ90%C/6$rcoc D3, PPBrJhC ޜrb7E(Z2F"V5c{1+(InNCBڹ%$~J%ɋD2¹з}Őu$ﶉ)e.6lZO7(:5%2ˇyAd8ۦla~l6je:{5u]!j{?$W9&BXDe'L%*T> Dy-@= QôcFa#g(A[(y&ŘZ"5p{{2[i7-7KoUol4glCnbwSm@0q H$ԸvfF:`i iؘ[2A%XL̥c: ȪQuJQu  tikAT3Bv Ʀ?k?R)":5JͤJ%&xC)"7&5&PvTvT|5UpﺇSx{gVe`kkٿ8 @x`z$FU`kiP O#R-%ki s&u"%L:Q 5ˆrWb(E^#kx&Vp2 6TR$`Lio+`0E٫4kKMĂ01EM "PB` " ),h$wZI]5 h]hjKčHݏpA^f2}c\9qF΋EID Yx]ă!$\J.P 5^ 0 fp ۿͭQ$!v4j% RaSĂfŏm$ ?B7? G^a{sw0P?oGƶ;[;}o; m&<f{v>|llmagzy잃tvw}ܞaOw:v~OIMoWݛΛ&vtԞ:3?|=;_v={lzo"2V9DZOݾ^+[uF Cb3.Mtn̆ɾe)yWي* .i.533| |7z}3}ox?] 'KEIlם'{gB}_vj%3(ve`mȗ2&S[6ڗ F>L/{/4~6|fqC\4f?jcyqp:fs72gafR88p8y&poni>| /;d.O~<ɲN[`F^0 Ygxb4S ί9ӛYoӓrP#M'Xnp|ץ)gNN`}N{ǾYL!4P;5F$pjCPfq2nӓtph6{Bn -:u!j n?Q^wV3z V:*Ff`Vrœ=+875B1Vn c$\p"V̨T+G 4p/ f-(86p˃E '1^;΃r"fP(m/Q2B@JGMazqF&+eF1e|v)cl,s|o,Ԁ K̕z.,]0犩EܝJLUT=;.u"FcB:L]O:T*UtM%'Uɻ-]fWɛd ivGY+ Hxe^`%2%,O$pt{QA!nT|%U8Mg#u^l7+#;>jbM6&VL’^BĹl* P)YJ(XIal(*RDG_4jQN-*t,|sLro}}}dH&;+BŔ*XGf!e/ p)IQPDKRЦ˔ h< (OV|e2A`ye,(9'hj*Re5#>ɘ ]mo7+};z@bUx ,6f0j"K^I,_q$㙖=FV'A[fê)E[':$Yk)֤mu&c\ c&1.>UCq k-r`cș`Au$:1b<\WEʵFI&o#8R-e_44I_ޱ=6z,9:2tV՜ j*#!" (]9!C$ ֥j3KUW^%pz;4iDKwVYYyZJpZ`*Ub `kgHM+}nkZ}7ܸ7k[4{JǸ^殻>9dk k 2J.JIE J(L(+|Rثjj,,㛤mN7oT.f&#J ^) j+hA b7-AxDBo/ڧ Do5gNF!}i9<\ycmwjmŘ߭:˓teKsqZ׮qZntdCxi~bwFq>bMF/}6w ;I,$IF2[nw4zń)%lX S"}掐a뾆3A_T8ɜmgf!={P5m+6SQs<3pnE+rѓ|Ըعgw%83[ s~i mWJ^#JJ[3K'%Gwb>iz\\b-?>Cg8Cߣ;􍫁KE.^dmG}? V7l ;0ʁxq-5ď~cԛ oǕm%*`R~eCЊ\H4_n9:*7lmҸГ {7W)w(hVSV?*jj1S'{Zs,: FxFۼŵgoj]w=wOS,^,~NSi"ɯٻ_ۆ/1Q=~>Hן;z;_˲L:=TWu,%hmr>wPmVj lPU _QEG{ذg=әg ?N so7~ ?8mV9`苐hj'gpj}}AG1HI8_?;=N[Ls6{3Y茘L*dRʫP*d)`K@ hI6 UsRAY*hVA U ~{dz9]@ZQ-5T[VE*#J1x $x{w ALlpEzn}NvMwtjOڟǠS.\ \.*EcȑDBZ' @L>e@tjf]̵Tx͇v1.\kW'q.F~[{k@Bslȹ+Z{Zh'Q`c/jk;8Magڹ_m'y+褏xZx뛛i{Eˍ9S8Y9?_JJŅDxͥ-`,b}\\t<^~>2_ |RpsQ锽^'a˅ި{XuTUt9j@YǨK4 気({Md5Ssk.{e5OVB=V[." 0bbbX#pG ^;8EN{V_sH(O 3:xn}լZ (T[Ͱ,ֺƹ= [*ZY+LgOJ(TuhMhwE1-r?bS"5[ 8;zA0\[kKϪtJ\[kKsmi-_miRh:ۥ?SYwUNdµ)JcH#څ]_m]ixza~R֘au"@8:|7kО[+iW[j|Vimj|s/W{I{sZzI_-r[m{1[s3槿vRlE ۧӉv<0\'x~r텈ٲrn0Rޞ/tL)DE AEoj,0NG,ȮX.!X+:BduH^' St{~Ip}'̘*ЉO2_VvNAHI,3Y>'yt,oQ~NzW%ʎk.~9yh*aKgZt :E Qڎ`=Og70tNIXa jZ3%'W5m+?[__-#!xHGQVSe]Ij/uZ_|͢W\e_}WQ]6兄]yI`Uk0-Zo0x$c0T:k]‚cr&䏡cAY_ ! [U^F+ǫ=h6,o2g/XXepۿ]sq}j罿|g;0}?c߿|x떊_1]wOG=(Dj tYg|QNbPGyԠ)*.?6\$C ` 'VL'1'rE(HWNSѿ0^:k>3G_HFU/{ %MUԑWU4f絛:&: bBǂiK vd"%&It̖֡"ބ$oj{8YpU 2L`N}T:g@͛|9ju݀L-h B:k8뼲b>&Ŕ%vGDl-/̾#Z<'NP%'2xS>$gO㱤//!gvo?l70 dƒ;JŵifrQ̔ξFĂKt[U1Œd;=fv3ÒF ZīpYNӯ. j )Վ,\zq~A-% vR &P~Uwga:-٥ͼ-ɰQVRrQ`e+eQDy{=v5QTzR\RꏕmOsmIXƥ@Rrcz%2HGKͷVdJOVHűhޭ"q=PIFI=IBR$phFZF5y4(tXꋖ3UG?If.Bù'Z6v7$Py;ERR,[)1V\$<@}j{Ѫbj(\dpI@~ .%}j$v4.o{C^“r7|KW"rYl#3hq-@1Y4Hf-=FX-KwƢٻ6n$WTpuWEU:%%HʰցkXh@{npSܫ1|7 \r&-4E\XZ8uS_ջ$lɈ2#qCċE< SrV8$Al|AN!$dX! lVو=?ϟ"\b]>hixc`")P5LjN[.t|f8ѕЎr@ kPLQC<qݳkAw$S 8"-+R`K!0[Urt\2Qik=aU ^rpBiL$CL*';bTļSoCG(o@̐nT Դ#\TK8 .~%Vi(d5BBlfbQSc -GLATSTշw-hYwV`%C"(M~B.fla]Gfȟ .HF23O2ffNghRTG?EGLtcC9ƸJXUߊi0">ڵ 00Q` oήHD[=fvXK)ĝjZa"!xB pq QU,nJMP8y"jWk9UoCe\sڀ4|@KqX[>S 3Q;rTf3Q.fw.l ϛ;<oo5 r[=[q}|v\~*uŵYdxp=#JxisrL<"UG?/Wk{ }9ӡjfɃ_`Fvym ,9?pIGvBz4Ku>t`))愷uީiV.Sn|Z: H[5?oTC2A9mѱqIi8y—) T[clJ 4eպqIKIZgĐlM 3>AH'ӿ;W˫Ye/f^*B?umoGv`C1 /lJ>8*M{?%_I+& s}iGY. n ʃaB6Nĉ;> &(%6 ]W7 aj~f႒VQնOxU?^uW/[p.GK9 k7Ir|s#f V05~< ewI :D B\}fW>{5?gx|oVU(pv&ylܙGWQtkoܿg罴={?}axHߎF&lV7;3Ly&=0 ۇ5K#ph8\iYg{ͨs.ͳ|恄`! rY\ώjZܽ.}gp֙`𾠿|ބІiu?&r~Nt_ pݥx7Lrv aI}a4b7gkH{`ҏh%={=AF,]u`tNYl3| ?(o<4ryu}@OOήf|u;O:U( . _ֿ](~wqko[]Hjƻ `׷o{6[PǝӘ2: A$E4&۪F޻t|[%XC,v\]<2;\c|y>6/G|۲^: Cھ}fn:uda {Cԭ㠍<<@< ']:8-n$[fͬȳl؋f8"kӓ8Us,Iwҏ3~]\cD*f}\6-.uXW~'PbR`f<flӷw2bٻF&'ߜx' e[4䥣ϡ#B-\M-ۢ̆l *n~ -J8 Ҭ|3`~w Sq_S8tz~6&BM8nhk#> 'qj8zoe ga(y*~XL?g -S?xx/ŵ(uԝvM{΁;XAoËۊ(XGg]oB]ެ'g@Ǔ(z+͝،Nd;uw_l2=әZ_{u$fsKip?`V~ a0rMt6 شoVU ϛYx4J|ñF'fB5E/ 0 "m` 4۲ &:Nnz n,цɸs|ܟ5il ɇkS;:@cj MCam?!nL-PxLjP&䜦jLӓw xua5F]:ۺ_>uwjCstPJ`GUn}v-;۝q Z:|=rCvOJC~Qʱ|qb휭uʅ02ycۘC0!)%z9$4ht7Fqߏ_ki2rwћ.~{ћL/hvڹus_ށ}ao?|d߉4&70ad 8yDxyi"qq'}ӻɿ ?5ra@9o(U rZ\5}.bt񫊔8/΋:ofi>pd΋@קƧ03VaϠdL?dGl'٠,X%ʥ2k/pUzCeei2o_=mXFh=DF㟇Lw[~//r} ~7oB [798TΦo_qo讞 Txa'}cB1/o4, *{EۭXmbE/͹/ⷤ?1#|?T.;(<>s_|5}wNz7r)z6x< !$D(:T|pz8,7N)~e>Ӊ5bXh](dm?Oo_&U`!.^ǣ1wߏLqKՓ~9W-?UR~<|*IZAR 2x 當MS`hvF YRU-6tzWTϘBȫlΙF {ѢBQVA1I\VD'yX`qgGoeAl^>QN92j޵P^.p›{~( 9h$9hWƹ:9 p2I~QeQQݻ;Q$qGi.O)uQhN^瞻KH~YШ9뇤e9#X4h:YzTI_^)#]DS#W8b Q=slO+R!gmsO}hbr2PINJfbMlRo@]s{)pL'jAohuFAv۽'6W2^FsEޓ $E_D"D$V { ʘX> hsѻDp(qC"BZUB5U>>;L.S'B$eb/rz0JD4maoWZd<8_}>`2Ws~ŌtE=Y|twHD&txL)!<_FWV .0@Ȃ(7ylFrfVnz\Eԭߙ WoPU^| x:Գj5<&#mtޝX7L$ 0lEM 'js&+sbZ:gWN~o*QuaFqz&h.scN LpJ҆,<7Z)YZQ"嵔>Gu2!+3W_= !۞_q萗)[9C6j/-^D SyRg4P !2E4JH2y 7}2tjM;U10GV'j+-3DU8izIU [$-'MRc1K-H*Riy8FRQ== š.zqQB H{BΘJ@BHQC23k`є3q%q#իw]ܭ g;0i,LDg(b5Eǂ" XB$\bA𛝄: kNok3付X=kS1fl髙"_'4^%[ÖscS2 &6%ZTԂ MTH`} w%bBIkȗlwK1"$mu s^vuz#K(.-+g:HuLԦK=(Hu0%:XI5.t|g 4: [Q]wK)@@T2]d$ xx2j+ϾE-X^d-]p=} @1؎J)ܙ @+QZR~ tC(/SP_ ͏ˎ{䤱 \14\_2URs0EX]7SbמS6uqi&!U!$=3`[[og/c;@"`loF[O㑂wt;U86{Ř5X9:o{syѮQwŮ)HӬ&"*0[\t"tjb8.naq(ZJjFkx;C8t,PDbaߓc5PG!ᚨ ?f~/Ps%'bC-@!:`j6&t_^``j.+a*>]Z0b繖t6)NYĢ,1X[^A-jiP58BwWj8Gy޿> $ H# -H!{)\ca2 B.Jo8ihWiE%h䉻I7 ҭ)bXu- ڟtk)9)EB(MQrVvz&plGo/v\o\ld:t^DEwNʹNQTB<;h^#'/ndcsbfab'y1hqB4d0>/7CtZ!+q7;/{?317K'-T6PQl(D\~ΓWa& t8I88)z޴ 89 í+*R?l Wz7vxȥ!BVK.>:9Zzb5o7V ,bFl}wvbW&+ >?şiW$!0GlfMuKdWI a3eԽ SjO58h~vQ v 즣=* 0rc+gu^1u$',T^(̤ *Jh4 3õ$#*EG `*IE6V.:`QFLX%yq%Y3AR㌓3Ѧ(OVZc|$"(fT%dPF.Ev1bں-@w{c!_.H-Fyq[J.n-$D9us9诠qZ\j!Q1~BcIP?& dP5YSE`|t.^|֒}Z "I a[>lHĄNR`D,#?n>yÆcHRE֎̟R2`FRj1:R =:c85g BIU#7]6 wVs\ ?MJO.@=%X "R/V~Tgzr22֟,reUôUե٬hOΦ8SЂ & S%*{}3Ԏ:#r?r:zMCd%}}XW-\נ]!܀HPR\ P T&ZHHjB!˯}s k?EPhBG*0.0]Jukd. RM5l|(b4`'Zdiujp5ngۈ 'å:/Q '4ybʼnFlg$0mWجLEfsr^JTmG#ISnRhm!_wAHq7/-4x[ɨL>/ zɌkM &-E\萬Ƭt&x!WvQQX=0x)jXm`2#f oFF8PѹG,:.7k,K]x6%un.zG9CR9k#1:չ]:lj29jplsF(i_z^7KTz\qm˥nSD;S"r)o2fL0KFrt͵0 齥:Tgɥ7ZKN (`Z#f99 RaQ4s!L%S۝|2jidH, @JLByN]f2 -6y5Ie1C:?2CZht5}ŵE55<;ԲuC,@JBÝwVQHLhr#к@kƩb-X vHwn/&igRh T98G%J#fm 2FшhRͳClGтRq$+bo/ǹbx2}x=$آTeIIf1NeIb<ɔ!1'T(]sWT~̑Hs^ުTlmR&Vֶf7W93Z3/}V @?ܽ |k%z$]7J赪eܢYrL7uGI:{nӊ\_ppr̕~A !j s%86\1`1VX{dB ]9^[P؀厎NԠ=n -}k_0`#qɔz^dH68ًUz5gX P7yBJRmH\D#}?QN@R }ػ):4K!cZ+˄*CN8g*fv!Q"T2b~|i1aqT+ Fj-U d~I3`0P>H㶩#Z5<pԏOKYYCR+(~t_{wl|Mqs^_Wqi?|U]_oTkAL+̅Ǎ,4,tM񾪼`F;'b`&˻Q'fƹ^baf3 6Y,@:'f \'ZZAʣkAm25#[ s ^ׄ%^+;|n 9Ɣ77?wsݏq1|6:j`ŸWlqˮN&N1F틏o(&⃽~{Op6Ϭf uWWN˧O]qaZ2JZI߱ްK6 "11!с @^8[WQ:F=@a. 1&ԙ痪~C~=Czwo;`]}%Θ?&/뛢>Q/XJ^Kѫ]U ~1,$* -DnU-|]'5F! 0Pn:V?[WZ2ݩ+YN876_ӗaXؤj4 i׹%b:ؗZJc3炏>͠IG}vLQD@P!LiOk߲$쟞\&HJ&JT]nŰ695L 31:s",׹ۏ|0Np"U;,)4J&[%ʃήQ Ǟ6:ٲBe|z0tJہyVaZv8\m fF!hIaJ ;>ZĽjm!ژ{܋ MFo6o^m>SG__ /#2|֗쏳}3n-_Ǘ+VQF[\~D P,XZݸs帴D(92ړ^jwP2w@)Xik󯳉5̹1#:q=|crFx:)>?OU׆tRjV&DAn8Ōi:>njcS>b$6L "^2nHH9/%@1! hÆ'*"L6BH):Ҍo]x\ *5U1i %- F6²%sw+0q3?K0󯓸qЖRTs"0u9W Saubld0s\s %*ݭl+fA"h"vvBL\ vƦ=ñvYPIQ?OL㡁 F=4 eIK4bI7KNKAs)8{.YL-41_dqr˖wĥ5 5kZ}ژ#W'EJ$x?9ccDv%:QWvy]ߟ?NyCG#W`k_i9 NhTMJ[ؕ1wY;U8r9iqNΰcu"z-b̡GT 6Uӱ~Gcb I N7)sA5Lgqn(pJLnXV1 ^,z  xVT 1`9NvՔw?>#ˎMz$x"B% $Au'I:H1!}ℊe;}b*ȓd:5=%nǏ_>]V>ϱE.{" `\x͚/.&h,'Wh]U*hi0b_.6>vx6FECDt"qEcȯ% HPT)s4aQ5ξثЬu:[^yH/%|0[yk/guLYQ4L'4Q[kRE7G*1*VB[ksC U)F=֮&>k YxeFbF(y()k\G/O: |=j4rfͻgn&"Hh/LUR΅Nc@EХ<3-L>}WBh>WfpX$LK5CY.9Ͳ3 PTOҎT]S=LiQnpt* "zr)'c  B0ÃMj@-|̙sdBbkQ vP3VBB#}<4id}Tcj jAeu䨲k ޠ10BĻ(!\@*V{S#n&kVmiDlGPqZ=*{ ^8ެAFL9:y G\2I3kL. RA>y1+"bV^I=`HKtPQ(k7.:F.n~M \Xpvt?7LPZ4c\bu|Yv`Nɡ ڏDL]!V0xZcIMyCLL,"DSV$孁:kx*yRWX$Glk5] +q¦V1JHYs!80w~oPCғyq{o;mrDHC>ԺQ1tFVʃ)6ed֭=Ѻ !\EKtZybݤbb ֭)S9m(Bj6Vqh݆АgG:%u*iWl)|-7`k.aƻηVכmu|! Vj0z~9/Hg5vsm&zDٙN$pF$)A1 @֗* F&bB5fE,F5Rh7l >s,/t83@ i8`6ӡ4>.GW88(i,fwTxƞψ\0:ƾf*ꜫ8i.7 )&>^v:(co`2Fp%?jɏ߂n(!u=dЁc!T%ϸ!K͔'}&ޅoR-4#P F |a)tfL1,wnl CnCec:CQKfN&nÛ2;7A6շ0|݀|v rLg7JۜUfymm MLM#3:?Ϙ2W mL~ߤԷm^,} w#e?X.j94?WFU}|3O]HMn=pśRaI0Oͧ^g\Q?zwk.Ԃ~d^ݻ'Z 7|K꺩kExLi{7&gël *T4IcҲJ'- +eM-5s_oW5}_zWS +X0#^\5t]*~6B0$DYG'3!W @LW֫[9TB Klbqv<Ubs? [TA({l+VJ:8Ltq;NLgg>3 `BvsM/c`]L ccj1ͧ4{l6g couHM_ɪ]AACJ>D-XDlӬk~{AkR{vۚs{}}?Ӽx FݽfkrD|ﴪ)"ǪSSaCtl|/H%vADN:# Q礛^#CLBك $/d3n:`iT3 h1 J'Un&5:hpJ7̻ps {,i\X@!4v5xBlIB t浶 bȵd^Zj- y9XvB`:>[{Bd!lXtKCk0xX!f-^ `QùaK2F+Ҩ bB>M3b))$4"Ku1G ʗ(4|;~?Jw߲Lk6Lhb" `w|RZl@z,3 NXpa **kUti}9o0R̷ BSDiDabo Ui 3U`P&u=ֺX"X5`Ymd\J в(\6u\ĠM3-hK2-_EB2}B^8r,WdDh!"[z, k1~v*MfOڣG+U~.6w7i-3.أdn"Q_r_z5mq5J*Id:[)QASzh{m+K/m{ 8$9n.JRl޻[O>r6HJecik'Ƴ!Kw.z{;Hd䰏aSfiM$P < 2[%)8f+U 鹨oF}#VBٸt.ytV9l["'*p 5PzSGEIa&Kda~#2WDI-$ةuท(dZ)$zd4R޲;TS[ϱD/95SnHכTgJжPlεwYԊjശ{mγRJxasZ)<+m\JM^Y 8RdެRyVۈ߽9`N+gK J̳Rnp(aBYiKp9n+.0-fb(J VTpЇ VT JJuѵ-D &!V^)vlFJvm)8-ћFpћ9Н;G`Bv1ɇ-/T]^4.Ck}$d?@{V3&$|/FI< -c Ffqm$;`S~$1ULI$6 dB>XL4/%Y" ߎ$w$MQTLƽehxid;}9124zɖ$S=&r̩9K H sƀ elbN>1uJM\MS QO(2(i0|ƫ_ 3w7L y~ I{;(_zoaEH}#=NVȎy 1]ȎF Ss` 0:{lӓkr;c`tk';ո[sj[>9΢ebt\I\2pvVW8A{4cٛ0$.;E ϱ!xs79$@ ٝګYW+6ȸ4u<V3hh5"x"ulnt)D\Z NTՀ2Yi&њ )AAgwEe!f˧WUIk\U%YK2**5s-A4z2b)D#ɿ!~R&D}DȊ{b6Kb#=iX( ˮ#R(UWh<~AGxI9kskDf) uVT Omdl>F&u$0,Ԧ֠cȳB:ڐl_d?ڒ ȼRSǽY)_FxMꬭ̾BU/P}IuN4AXiT I︭TP jR&J|jJJ'CaXi^wOVzV*JQp}VZ@5!;률C,-Si]͛Z8z a4ZAtaa| FA6H'TBNs~D~6#2oWXQ~v60'ӱ~+꧂7_8}daqo`% Ymu3و ̯z^h۟ !l;J5h2•@&&W.l _D#$m_82 BbXi͡2 hLU;""JFԜA삄DP'p A-<馗nPǞ(䱨=LB֦nC~ x}G"&sv$lm"'B9#ibu]@Lc܅F3dV9ɎȔXH6=FʒU{SֱRӖ!X33V7 Ѧ\PfJE`)wAaa Ѿb&+ K3ffy<Bf 1{jj*5fΝZ1 UF>C5b)*43zR"D`Juh0w,qyx2YIM(L6C%(k.:NE$8P 3T9V՜ +8d_x*[hΉMCh"׮OBLkBt677!גyYkF5Y鹅=Hpsi]x}; >uzMJ#A|̳:qWQt`>>A5u)e1aP$)6: 02LMwyq^Ç_gʚ_amgl z;':P(Tvբ XH (A1cRUD&yy ,qJ)ogofqPH Dwqj̋ )=f߫ſ>ŻJQ9(@Sx[yysdd"=E'_KB_5Ɔ_J⊦ה7 .;˿i'QWTv>M8JGyܠI=;G;~NhHb s71tLti(* ))U(K6DSKQ $*SJ z_(: j.^I]]u.C c_>:f( hFk9miDATHtE+<҅2R/hꠉj}[~ 5)7„’M*Pϳ nAGjP~w﹄#H%ۻ@ɲ~7 W?/@@L 6ogf"L~7^,-;~\___|:3ګ|OO\"34LDz16w[t}~،(d?dE͍WXر6J F_I،A![xTP?go<{Cp(D#}ŤEHBiپ_^{ IDݲ{}ٙ篃. 0Wyϵl͓:iPQ1ewyIN'<22Dܗf.m8p3䍈\4wM2}Z_B8vq5. oVcK8<`Ȩ{%y22͸4.ڮ5̜ ـyhO*{Ϝ P('ձ`WBijf=j|s?=(<=?ZF!5r{k+'kvOl@zO>XKI3;xC5bZ5VswNj58JLk~'yy\<$`x=~?:cp~[fkbmImEK7}.,T^4:yzo's?Ʈn}xy9w߾\+yQ[{]e­֋PX3rMwtַܖ0ꇭ?U$䕋hL卝0v Jt2F+<|8햝:W.92@y9h\ bD')m<o#[EL]EO VXw d2zD"JMGMKJ V0JjAld-& u&)6I1o8&G6usvTa&P6"4+:h:*aWWS*KxȒXrCh8 K/!K%PjS4cNhpv㮱 SΠd,BpcL*uߗ>$Xxe/% \ > V%L+V)ȦbZI!)-UTLO|1![ymX|͘lĐzqH ƛVz%|FLCժd0X;bd2[PXd)P* @ C)TYO–]loW^xrU%Fn2_n׫KeJZݾ?ޟԻMkT^NOt I-y5-uQ X/.}۠9//8R|X ?)JL9erڞJM o[W5ϙ!Xs  5#2&*"X,NrK V8-4%<0 ,򪆑-ήV( JeEJ0*g7` 8:rz`cQI 7՛[%d3.@P ?oV_h}$OX4a#2PDy3]Ts,J€-O5xE?];٭ .q"Fc镽aS NL0N0@'qwcy5ewЃxv uXtsgK]NWbɆϙE+@)oZ-7߽F&&N%g2o5+͵<_HmݫIᆑfxt>c5EQUcaNb]X\#Ӭ"~mBP:&* qWZj,%(4f# 4,r,l4l!%TgÙlC`UF 4Z-) $+9Q[Z.(){cdLQP/FTEYI!Ҹ n8ך0%Hb PY%pFHSY+?(hwk߮E ZpoW \ObiWlpvo-O׋:ly}aX/kw/w   l}]\ͧ(l;䬗po {Zc)=KC"b? =6:y$A6בH|ڒ1e=Ii&„PF`eȗ!w豭+r`Z 5<` )G牚~>j$q|K`H1q,=Zh9>oH!C#j}8q%Q3Hw/c?j]Ȣwzyq^vķZfEGϼ_l^+`L7:mYygXg{n&yzٖ: 7Q,iwLTpE!yRv!.>=e>ZcH938X*@rAkf[pSmw̸mw̨Ou}+~#fb8\"t\& Sj "Za "S4֘JR*UfE%6`.ϙB ^@]aaE6:]7a%%Yl!nOL5 j ^FIEvp{O-e9F{䚝pH"d»"Мx nn{]!rzTJfJm~ DBQ!XA7v_އsѪkS6=w[aA5矁6ҳW{[ɭ$*t$!\Dsd M1ՒNyXD[H!&D$@ CkL-2z? $Yƃo#IT6*{P-OZ8O~ޤ=~p\h*^};}ᾩۿ'82)ox^b\y1o8:/1i >5=6n[VXpZ6 g*Q0Q:y(FHg)i셟Ԙip/@cv.F!ި褕# Ty}&{* -eK /DJ2 F@Ҩ*i`^P[bMw۩-ZZ'ݚhͦ&3*Mz*e\kZ(Ԟ{lFemg!薑\1v}9Oi-z/5_ŒC/FfPt,AzFV'!{vMTlޢ+(jI7\oV!=%?fB1k:upooGm/-΁)EMcEP2(S"=a%J&ΪBm$MLrEN,!Wί(8OLSUȥ 8Kzx $n)JPmY+T7OdtTQ}QnjAЪAh"8w@g@%zT:y}S s Gz)7Xq|T" $%)`Oð}v]*!]^^6sFw_ 'Ӗ8!T3B}B/PN]ɠZq\.p@G &2)9SI`9g.$䕋hL5zT~݀` FtrF<}I 3"[E4KyzB~xlD햋A~#EHF/[~eBvBB^Td*.(YqWϏWļhqxObkټW7¼3|j[B?bz/!{nNܛ#دÿVo'{pe "ޏ_ܗ\+'^u͗:6qK:XŒ@v,x GXp%Pfp_h6.%cptM]XS@O,[c҈W:Yk_aȏpfǘzsb!d٫d# ȳ}"$%Uny&?'߄'v*$7O65in78"$MIxI\x]d?]mo8+F V=,Vb8{4HI6c;; _Jvo6)LOc˭Ubl%gߧ) - ?U oGgqZ) mVA= ߦFa6Z~_~%;VVg feWW V$۰5&/ .&k!u `kDЯ:FGHwnW M=Z]xmWbae@!]3ogdNEZ(/}>Liku#6 Yr2}TX]|0SJ'O$K;$QtQ)ЍmK\hFz1ڱB5hmUH:w^,#/ӔvL!F9K0_?h5_Kݼ|.rdfpb6sڕ7Ԯs)kP84*@O^ucoař&0&@-ED蛯z0!S4=;p%jی! #L[J;-+"m*gUGo0U;&ޖ/ֆRfTY9ZDՄ;x~f;._YHܯzڛf4ӧOJ>fZ \Rm(&j\ )RX Wo#yC![ 9ސLreY0NS!/ǦȿC!*X;y)ybQ>rTu{s>Ѕf#]p]6d۾5R $I2HVn$3`QWQM iy4R&>(e%utLX1cZ4sH,haZQ%T+*71j)B]+01DBmt'!^H4F:B5dLT)<[Or=R,m~Z7s[ju1NaK׊:v{`ZʶVNONΒ=P땶멵Q$*$ڽ3+ۣoUS::P#%x'a%ȡH"&Mx#ih y:Ȉ"& )hBVXKO G&^z2"V1c#釠R Bv>` "QD{ nfk6‰n03Cq6 ؋ZKz%K pRWeҹƞ? _Ӥqybhc,_W2a(ryN͟hPk0S&`y ՆyL- ծVG{lٵxD2eR-(iiG"-02d8j³5~ 1V9ǾĂ>5W$Ƈ $߸)},A}Gw[0J:62ݦ߸)#?n<A^kq(.23vkMtǦ2kN6e~%N_* K7gs-H*޵ yrzlN w?]r{I߿Bh ~S޵#㒭'_ܴ$v.}on>yÛH37ڄ1iV"a:xnSxP7C;>}aARo,uw{Zn7O8p~{Pd/i*K.FTuEbm2,N[^OހC%ͼC5}]ζk%~[Vaf٠e݃M gHK- d6y*.Ý}JTʐAP\^-/)VJJrNci(Z.iA1 Z,c, ˷** ϣ1TpybH\}5׷&? B]"mL-)}զw mUyǫ=b6쥅nH. UGtBC's ݤE4\xv ,@&`Xi9<˱ɜ u@E0ֵ6Dt49$l^8A3I>2Sre1Yؾ2֊F9pob XΒ&ñR0ev:hqe V{ 68 $Pldݼ|!rm{ubNΑ.l *5nUYr0&i4 lY'K H8F [Xq&\J7CF+jgp*spji4vKןUzKLz%0abA*9AGM 9p0a" dK%  Bֽ5ab <@Xu-Jh,"bDijuw3)0n4ՠ-XT.(EiA:Mi4]i.$ Y94fHeOzI)!ҷk970yZiIAr=(L jcoB|UA`6u^o]ハ&1gЋM-_nSZo>*3.Ox&^`ƱUPfy)QhW4*[ۼ_>Hnޠh7mCruw7ݥM^Mwle?ۏ~iwLOQ?'kl^,rk/??fnmIV gIGymV Ec,IZ~3B,y` e40A 'FJɖja1EJ*8p]`-L6">Xaviw:Kt T*Wa/i/E W+d+øJҗQmjmyͬ:_kôv5jVzV׾wbb/P^TeRfcYX[J/JIwϭ2f&ջ_ [IE'?O?1G4c)KVh O [Y ZJp͖-pBV + q;(r_w?'%;PUֹXqpU8lEu瞤yx)4r&UGt@Oj?/<VOi3w%vMOmKl9 LLZ+MmGVlǨpC=v+)Gл%SKؖdĠj$!F>}И`klg9#"@iJŹkG)V(+oPOgcy6FVJ+im΀vl X4 gG/E|p3ÅM^?G-ztù5)eV)Aӣg,2&A quQI9ţG7jpJr2V HOM~6%+sz1:pK_*j$=:p(U@U{R9&u`ALC4(Ȝ)@F mN7%ԷA[{]Hf{I+*J[inkJe[)f"d۞m`BncKyy|Id׸X"\v*>]/r[8󬔸;E> +} j&Z[)d;;m;؝C42/JY)n?+owQVGp+Eg(R˳RyVڡ3W+D+%gԝL⳰RRyVJɄ/l+U+V|e,Tej\Tfzu+.04ALVo凝DOI۠DvJXaAR%i= j=%Alf,d#H b\Rx}VrT"f] SAZD® tH+rZ֕TG*e^b\F ,f"]u4LG$OCfn4 CXH,ȭ"~Kc=ByNqHH&_>>{mo~2aJK)`$'n'L"5 D.jlڿ6: \<[I .ز}x8_ a-jc# Zk 0Y`q1j1[ aLS-];L!JH-{Li> bZ :[#"g Bf0\!p.Ch cPF3f,% "h2[Xq&\P!4ztcB緜̡YէGOv@Ah"!t &dr3Q8ںt+fUㄖQ 4(| 7S;~.j!tw2TEvbšRyܡ&pW+D+C590x/r\Ӕ^ͨf u=+Jå*BZV˪ļxqP]/2J9+͇p[LgO[9 H?Լ#$! Xfs2xJ E!}2#]ZV )xW^)BF,WOрH䆹z"ANd$r[ GѠowN6rIʃSCƜ%% )PBEBd"TMXBiS\uy=IzmxT_}E{OO;e02YA1|sJ Z-B*g9dQ[(we"ewH˾s(UbfAҠ_>jy|' {wWȘ:G2ȭ+(Je/*Kkt+*VTfǀt \INrٌvi"0y6DR\A j~+3Tږޅ9[REN`eHa du ۄw߇f6+Cړ>Vy^'=oah(߿w96#'з+YWd?}g?3&fs_9dq\NI!/_|UO{|5czUY.o/M[SZ֯? _o_={m<[w՝_H$__^ۄiY2VFؾpY&S|SK ߀{Z$dGi|;J?I"0yp:ulPк(> ߞ\v ]{\y2a-tAK!Ӈ4sbLvCXP>tT"QT)/O>L`ggowC60 ;(c}ߝj6'o!ﶩ&l\|hGhW+Hu5ϧ*"쥵={{7oV^U^^`Us b;?}k ma(ǩzE=PgW}/ [FeK<' y&eSRN|(N tr&N[C{HևpͲ)#O>n )x\ĘN7R۔YQFnKޭ y&٦ Oe2As\rzUsȉ&[&NQܨD;Q_ YV7OW<u'f!ԝOCH;)vc'uwIG~2zN69 vҨsvRp4c-Q sbzk;ڷ}oЗώ'`́ѤeI AND `4%nlyÁ,yq \!0R9J[0֕e(y]\]t^2v%O $WoP-#k˽%ͮ.TWvF{SТj9{D~^|ȩ.k$6S%/: ':CpdvU"J2JĮZtkJ-E9J.Jޜj% &dl2S2 9Nj\#j?i\<W9{0N r\|(*ᬦ"!ƙF,fT0ǣfvЏ}0S I@1&o59[Wǩo_mQ0>aYWAԶzS9Ns[~(%鑀_djl(W؛H {S6wWQϫG7Jm } of?%`L}#s.xq57™*D[-'9Y]@i)GU 0VQᤨ,GaU劭p942<7nhd HcnRrՅU!t)m,` F늁JYbaC"ЄfymKD V̥T܌3Ӣ?Ϟ⤘w@9nN3Dl|fWmXca_2,Iskװ1vrr+˗@3Aqz";S a|1grͧFw˛?WO;,:{ӣq?tX$iLAKo7yIpd}/Q8)f99~2OOkZC.9n{{<{H=N0Jl;GLh%-=2"PsF¨[;jrBg~ YH8(W[Ba1o!/x#9mSD'{wח/&_ׅ1bm*U;xUTm9.juS>h ;‹3uMVFʄ%ҭjƍ3y㬘#kp7:SΣ#0/#vύRA@'"#@r`5舸MZ,cUgQpekxN?]6};緯Plg>sZPDR@w>pHP&%a,{ uшwKH4gylRD<s<(Fp`'b2x4992e6|L Cd JL7"c@0eJx|xJx)5G#G{<'hch@fw@N[Smd]Œ$Ȕf04Si@NH׾iGo]}+E@2?|*Ӎ?5% vF6pbSz\8u ik䋊EE8M)Ԛ5RzdA0#،|p?΃\pcۢ N4ɪJZYQCaf$ճ"'/dm@*yQP˓Ra'"kЉRԁo"AvJǫ%@pNn칹$}h#AD kALC2 iN' a^RN![ %<(%ɩ#m^ u:Yٷ_oC[# p{nc.YcЅ|㬼[xՔW/f{|L}04߷Msjqy3#(kd JrMIQ2( QkúrۊP+u .g¥Lsّ%2ܺ okG}5sYg|:(ҚiH2,c՘-m32syFk$VEQKbܯ@'*uE(]3DU&TJe 9bDeZhCc>L41gۏ x0IDܿAZj܂\p 5qs>2c1V~ў07IhkhZH$'%S<1Y༇aPn2)RX+&j֥Vkr-aEfI<kn i@ ΑAd0,|B)0xgq1Nc5-;8䆑.@Q̑DHUryscozu ܧb< 4[r۩QQ䖢-]䣁VOd䖢Q_&xZ-AʇZ[6ٕ֔&|:b q0!T+yەDx+}_ $p001eEg [ǔI٪ϡĞ +,䅛hMqv;w>z\ĘN7RtїiRph_?nY6N?n [ bL')m,7V?һa!/Dl*r*;W%΢cE1oY.XԂ y(`U=]^E[̪ww J͔)<8NmjA^z)aː>lLj52 {T:gJ%`U*q2*ʗؔJ j30vS.da:#7 &47ul!}o=YB3??ʇ umߜo}[ЇT3Cᒨ >vM=;b^-l|7(VlE߆!Iޖ5I-,șuyqǸBDƊrv岏IU*@#C GBzQ7EpWtҍ䌷Fm=8 o6<`$x*&c W]*jb@5Y)+meDi}0s5aJ[s7+|DŜ n)v95 wCW mmwiêJ\X R^RaCŎ"PDMiR/[vB9J2RƳwm?m,M_AyY6/G=iveH`rod0x6$(ɸu骮$n/zg9[&c07NL[ -c{C$jXeGwcW8G.rs%ng |V$d0uC{}Dڍ׍a}=q\n\^N4zq0`6ļCqFͱu\vPЮprr|F;ϗ} r4k39$v:Ǚ_n'? n\ôڋqs.oW7Ob-ȟ-WONDLGlېh?f~¦z}3re& Nվ\Юiu><0`ͽ,{I4?U-$3gX@(Ü6jcKЖG|RwZ7] I>DXC  gdcZ.R  uNaIGڸZl!G)R5H-霒omڋomSh2ͽ_Ωһ7*|O&?uUx!f 4Z:$nĻ.8FGlgyY26m_z"-&J0V#GiM]5 ;RJɅB,l,1: 0Nq'זWa}q/ :~ {tɍS!|}RV3N*E+ςu*Cc>FkM*HabS%$\H*P!K1(QjI6RIs6RSbb+D۪AT1Dݩ 9G||HZX]+Ω==w2a7*}U] H{{V' {ueB:vnrbT9zc"˜2ZW*8w<-ϗc@qe03huQGr;кwѿ"q?kLNpw,S5.S;Ni%LJ> asΓArI"`AKo*`!'†F.1<L& 4odEBQ` )WVg(۳Δ=phǒ^jpWlyLrAJ{ {2H n;='ϕ'|剱{,' ݖK~z!59\-܆1vӼ &;4l%[ZɈ?;> P >Hs$}А\EWX7ѽpHӁn 6-3;XCfbm UtEtN)ۆvҔ7m=Jn?(͖i>l#j%ףZڪye֞ܪNSK}"C'IX#N[w:"ة1H5x^H7;:C@vBbx4@`}-µ"6ZYTl2l/3x"b9x˼S9 B&=Z$wlDfKAlp|_e4w|(J6`+lDEmZ a{Zt iDES(VyA%D(%]6&/*RZW?nv Ql ϱ+JA"Oϡ,l 3ц*{tEL9ev6UiMo(;WǠw0V[N rm6_nrYLl2|Iilp2F2K/7}n> l#l}XpP8G.؅gJ`M_;m'Ap{Ci/}3\v܂uk^#׏cm 4q jv|/k3a Y3dt-h$[O91hy;A鍇+P^|gk+$EnB }{* 6NW8]O;.ƿID`Zd|8l|x2&XjN?8>>nÖb Jc|l-9ψ^JKɥR|bZ5v.(qIe~&ƨJ | 6AJn.ATH=drC-[G|ΎGApd S>ەl\rC \]RHe3ou& Y0ł!&GԪT;1$Vi+=vãBӻ%#)2IOflJ: bȊ#I\Kdk^?/^czD#aV~EWFXkS>9XǥEIou 3J/9g89pʁC! T(eQIs) l ic-of ruNj4 ۉ"nـRup][m()jHp }&pNHoo֚_sJ|KmBif~R)8Βs|͌,0uxi;g/_6{,|%=?\pb\iNkC#9PBYFS݊/;^6yĘ,j60G`$,,\cѮzg{rv_XŠ =d=K3y{E e@̥zDL_?l68 DLT*E;YF٫`VHxn|pʏf(y>ݠC`.01BSP(EǗgҍDa0řl&b>XuCed2{\(F~jV k| Q;۠j,|Q0o")(DkرȷdhgO@.DD yD1B }x`kYh:sq,$kM{ +(і-=xWLwSH2!eQle\{p[$ӉEYH9Q( nCgɷVβ/O_]"h;O~2 BU._瓀"f;=zkb1Vl~P,CI;tXm`[[뉊kmXn6io՗~$yWE$98}g(iDP$-RfiWU]TPR "~q2h*W s/_׀\ hBbs]&h*UW.^F?N˽۷ ӧ;>UCNufoze1 vMKEdU@kYז 2~>ԓV*u"\Kg,ra%RKgFYœmE0g+#@&QPΔN;|8;\Ȣw;an,VHPhT~`p"5,`PaCrR"ߊ&Y[Ʋ}7e 0,d)f0l,aXRJ]6V~|u0T.5k*iՁDUd^Jm'ػEFRze4zQUXJt%t.iͰXdDUQpWzq_f?kQr_5@盩z.vw O`P57WsevCn^[9ٯ6JЖvgnA[4QWanCO0];re|tFjeh43F0?uƧ A'F?TnA5>~[IȍO_N\r|u%?vWh=,F~x ^+o,4Zr@aYR^Y]PXJ5֏~lcp;U8j# )gSLց+bN+Y[-|A*Mf^ބ f)p4Wf 6Rm Nlhy Dj VhCTR#xJ,jVX'׸=Z5$K @XH$AW1$&hzPØZϚ>ͶF M3FԌ3|)"J(E8c9 bH R*?1%Z杠L Kk5JskXSp`Y*C!ˠ-!/NZ0Q(rt8i)P;zfb`)#%4˳ 3:-H=*8r;PW֤ZI[dʿC3T`[ LZIBَan6K`҂ ?uJL*+C'o1JB$ckIlNkUMQ~χ[dkUr&;y$& I*iibyNZ16{J,c +% su"KGl7PqJNJ-wB Nѓ[m\oyX ݎ6Q}4K5om2h)j+!Qm6nkl6$T҇>]R޲Azym-҇>S}FK.ﴮ͛VjN3ݾKi'64\\67` 3PIU?KvOK3!VOݮ,ݶ:yS$݇iqqxxRmHk߲f Z&q7tgjX!Q_ץ!ɺ>뾮,oGgs_ Y8nMV"p } #P|cGy@w1׍kv]|.k޸z׎/L~)!X7?]]χG_R,3RFLN3?=}_x d^:q]0ABd- B'ty?mvG搙 Ұ$g iU8SFx|L)J= |?%yT1BĜ FAk%9hJ) fu[C"T>htra, b]s%zf&KiGoRY}&'^;q֞~,D!nXzm5t.oCLx2U_;UF+ ݮVnѯG(E& (wvA?ퟛ"X$XuAk{oWz^zSI7; <%8CP"%/us f=^e\J{jNڊQ}tjdN-#.h8IJh)jgPh.P y"g/P9Y2UTsN 1e9  XW~9c?{>O >1'3O+=rinegnNyu@S46@shgkd{ 7wBؤ9NϢ得`] L'AMVIr 6H>XalP闔08 Y<;EY!j҇6n_Pl[VK\'tIiٯv'޾s[uCc !J&G dvF%L6qJFꨥM?p$ۻFn$WfxXK 2-AF 5`MD~bEZ܌O9u^,>>_'"ɮl!xQǷ͘M>YCqY?z=-U%l-A&ZJ2zJ4(=ڻHlsǣ'  Q͘ 4˕5c{"јC\ӹ'h€M;;/qSHZ)tڇ=XP.r!,01@Cgծn3p*ʨjO"Uu<-Y@{;)wO*ߓ<-+f /pV+~:j?t}k-Í<}OhtVp2~[F_QvjK.vNdjeDGmrU_?W.:%)9{e(0v֭+BT;X!Y ̺u>Z>4/5 bE=Ps=Ws+]tux,Q}WkE C!,TVT;߫^`V+9TթmZzZ &LK8-BN5i)+h)#aZZPm+ZzZJ\Yyah ӅkaZ* _ ՠpZѹx+szT4\>Y|5T+[QKJ۪t̹wa:Ў|u@Q߉^ ߐ7O7D3 JܔX?՚2:ڻz$K呔ɽم[OJ%)@y`q7% 5c&kṮu_o$ Ä9Js{ pM$]AKi(MK6%sQq~ZJ‰ZQF{ y[)(BVAã7uWzy +8}zOxT{;v5GV(vݵmFR by{x<M+*[VN|r K:v[vKuYUZt^u'v祷^Z\/3 gajoF;I=L@#X5xͿjϪ@Y)>qCTթvE»8U_xvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003626670115157214670017720 0ustar rootrootMar 20 08:56:56 crc systemd[1]: Starting Kubernetes Kubelet... Mar 20 08:56:56 crc restorecon[4748]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:56 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:57 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 20 08:56:58 crc restorecon[4748]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 20 08:56:59 crc kubenswrapper[4858]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:56:59 crc kubenswrapper[4858]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 20 08:56:59 crc kubenswrapper[4858]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:56:59 crc kubenswrapper[4858]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:56:59 crc kubenswrapper[4858]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 20 08:56:59 crc kubenswrapper[4858]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.822873 4858 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829485 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829509 4858 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829516 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829523 4858 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829531 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829538 4858 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829545 4858 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829552 4858 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829557 4858 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829563 4858 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829569 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829574 4858 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829580 4858 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829585 4858 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829590 4858 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829595 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829600 4858 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829606 4858 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829611 4858 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829616 4858 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829621 4858 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829626 4858 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829632 4858 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829637 4858 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829642 4858 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829647 4858 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829653 4858 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829659 4858 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829672 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829678 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829683 4858 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829688 4858 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829695 4858 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829702 4858 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829707 4858 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829712 4858 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829717 4858 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829722 4858 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829728 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829734 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829739 4858 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829745 4858 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829750 4858 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829755 4858 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829760 4858 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829766 4858 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829771 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829776 4858 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829781 4858 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829786 4858 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829793 4858 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829800 4858 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829806 4858 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829812 4858 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829818 4858 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829824 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829829 4858 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829835 4858 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829841 4858 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829847 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829853 4858 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829858 4858 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829865 4858 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829871 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829876 4858 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829881 4858 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829887 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829892 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829897 4858 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829902 4858 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.829908 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830001 4858 flags.go:64] FLAG: --address="0.0.0.0" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830012 4858 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830022 4858 flags.go:64] FLAG: --anonymous-auth="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830030 4858 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830038 4858 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830044 4858 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830052 4858 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830059 4858 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830065 4858 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830072 4858 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830078 4858 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830086 4858 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830092 4858 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830099 4858 flags.go:64] FLAG: --cgroup-root="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830105 4858 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830111 4858 flags.go:64] FLAG: --client-ca-file="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830117 4858 flags.go:64] FLAG: --cloud-config="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830123 4858 flags.go:64] FLAG: --cloud-provider="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830130 4858 flags.go:64] FLAG: --cluster-dns="[]" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830137 4858 flags.go:64] FLAG: --cluster-domain="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830143 4858 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830149 4858 flags.go:64] FLAG: --config-dir="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830155 4858 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830162 4858 flags.go:64] FLAG: --container-log-max-files="5" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830169 4858 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830175 4858 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830182 4858 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830189 4858 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830195 4858 flags.go:64] FLAG: --contention-profiling="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830201 4858 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830207 4858 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830213 4858 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830219 4858 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830226 4858 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830232 4858 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830238 4858 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830244 4858 flags.go:64] FLAG: --enable-load-reader="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830251 4858 flags.go:64] FLAG: --enable-server="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830258 4858 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830268 4858 flags.go:64] FLAG: --event-burst="100" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830276 4858 flags.go:64] FLAG: --event-qps="50" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830282 4858 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830288 4858 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830294 4858 flags.go:64] FLAG: --eviction-hard="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830300 4858 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830307 4858 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830332 4858 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830339 4858 flags.go:64] FLAG: --eviction-soft="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830345 4858 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830351 4858 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830359 4858 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830365 4858 flags.go:64] FLAG: --experimental-mounter-path="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830371 4858 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830377 4858 flags.go:64] FLAG: --fail-swap-on="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830383 4858 flags.go:64] FLAG: --feature-gates="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830390 4858 flags.go:64] FLAG: --file-check-frequency="20s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830397 4858 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830403 4858 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830409 4858 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830415 4858 flags.go:64] FLAG: --healthz-port="10248" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830421 4858 flags.go:64] FLAG: --help="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830428 4858 flags.go:64] FLAG: --hostname-override="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830433 4858 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830439 4858 flags.go:64] FLAG: --http-check-frequency="20s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830445 4858 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830451 4858 flags.go:64] FLAG: --image-credential-provider-config="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830457 4858 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830464 4858 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830470 4858 flags.go:64] FLAG: --image-service-endpoint="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830476 4858 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830482 4858 flags.go:64] FLAG: --kube-api-burst="100" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830488 4858 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830494 4858 flags.go:64] FLAG: --kube-api-qps="50" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830500 4858 flags.go:64] FLAG: --kube-reserved="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830506 4858 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830512 4858 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830518 4858 flags.go:64] FLAG: --kubelet-cgroups="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830524 4858 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830530 4858 flags.go:64] FLAG: --lock-file="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830536 4858 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830542 4858 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830548 4858 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830564 4858 flags.go:64] FLAG: --log-json-split-stream="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830571 4858 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830577 4858 flags.go:64] FLAG: --log-text-split-stream="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830583 4858 flags.go:64] FLAG: --logging-format="text" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830589 4858 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830596 4858 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830602 4858 flags.go:64] FLAG: --manifest-url="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830607 4858 flags.go:64] FLAG: --manifest-url-header="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830629 4858 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830635 4858 flags.go:64] FLAG: --max-open-files="1000000" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830642 4858 flags.go:64] FLAG: --max-pods="110" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830648 4858 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830666 4858 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830672 4858 flags.go:64] FLAG: --memory-manager-policy="None" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830678 4858 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830684 4858 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830691 4858 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830697 4858 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830709 4858 flags.go:64] FLAG: --node-status-max-images="50" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830715 4858 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830721 4858 flags.go:64] FLAG: --oom-score-adj="-999" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830728 4858 flags.go:64] FLAG: --pod-cidr="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830743 4858 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830752 4858 flags.go:64] FLAG: --pod-manifest-path="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830758 4858 flags.go:64] FLAG: --pod-max-pids="-1" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830764 4858 flags.go:64] FLAG: --pods-per-core="0" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830770 4858 flags.go:64] FLAG: --port="10250" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830785 4858 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830791 4858 flags.go:64] FLAG: --provider-id="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830797 4858 flags.go:64] FLAG: --qos-reserved="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830803 4858 flags.go:64] FLAG: --read-only-port="10255" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830810 4858 flags.go:64] FLAG: --register-node="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830817 4858 flags.go:64] FLAG: --register-schedulable="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830823 4858 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830832 4858 flags.go:64] FLAG: --registry-burst="10" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830838 4858 flags.go:64] FLAG: --registry-qps="5" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830845 4858 flags.go:64] FLAG: --reserved-cpus="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830851 4858 flags.go:64] FLAG: --reserved-memory="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830859 4858 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830865 4858 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830871 4858 flags.go:64] FLAG: --rotate-certificates="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830877 4858 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830883 4858 flags.go:64] FLAG: --runonce="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830889 4858 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830895 4858 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830901 4858 flags.go:64] FLAG: --seccomp-default="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830907 4858 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830914 4858 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830920 4858 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830926 4858 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830933 4858 flags.go:64] FLAG: --storage-driver-password="root" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830938 4858 flags.go:64] FLAG: --storage-driver-secure="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830945 4858 flags.go:64] FLAG: --storage-driver-table="stats" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830951 4858 flags.go:64] FLAG: --storage-driver-user="root" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830956 4858 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830962 4858 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830969 4858 flags.go:64] FLAG: --system-cgroups="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830975 4858 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830984 4858 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830990 4858 flags.go:64] FLAG: --tls-cert-file="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.830996 4858 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831004 4858 flags.go:64] FLAG: --tls-min-version="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831010 4858 flags.go:64] FLAG: --tls-private-key-file="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831015 4858 flags.go:64] FLAG: --topology-manager-policy="none" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831021 4858 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831028 4858 flags.go:64] FLAG: --topology-manager-scope="container" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831034 4858 flags.go:64] FLAG: --v="2" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831042 4858 flags.go:64] FLAG: --version="false" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831049 4858 flags.go:64] FLAG: --vmodule="" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831056 4858 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.831063 4858 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831194 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831211 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831218 4858 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831224 4858 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831230 4858 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831235 4858 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831240 4858 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831247 4858 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831252 4858 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831257 4858 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831262 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831268 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831274 4858 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831279 4858 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831284 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831289 4858 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831294 4858 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831300 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831305 4858 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831327 4858 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831341 4858 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831346 4858 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831351 4858 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831357 4858 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831362 4858 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831367 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831373 4858 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831378 4858 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831384 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831391 4858 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831397 4858 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831403 4858 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831408 4858 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831415 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831462 4858 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831468 4858 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831473 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831478 4858 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831486 4858 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831492 4858 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831498 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831503 4858 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831511 4858 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831518 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831523 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831529 4858 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831534 4858 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831539 4858 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831544 4858 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831549 4858 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831554 4858 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831560 4858 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831565 4858 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831570 4858 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831575 4858 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831580 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831586 4858 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831592 4858 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831597 4858 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831603 4858 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831608 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831613 4858 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831618 4858 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831625 4858 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831632 4858 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831637 4858 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831644 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831650 4858 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831655 4858 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831663 4858 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.831669 4858 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.832961 4858 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.842110 4858 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.842133 4858 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842192 4858 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842199 4858 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842203 4858 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842207 4858 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842211 4858 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842215 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842218 4858 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842222 4858 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842225 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842229 4858 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842232 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842237 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842240 4858 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842244 4858 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842248 4858 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842253 4858 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842259 4858 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842265 4858 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842270 4858 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842274 4858 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842280 4858 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842284 4858 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842288 4858 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842292 4858 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842295 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842299 4858 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842302 4858 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842306 4858 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842312 4858 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842328 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842332 4858 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842335 4858 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842339 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842343 4858 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842347 4858 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842350 4858 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842354 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842358 4858 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842361 4858 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842366 4858 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842370 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842374 4858 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842377 4858 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842381 4858 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842384 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842388 4858 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842391 4858 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842395 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842399 4858 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842402 4858 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842406 4858 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842409 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842413 4858 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842418 4858 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842422 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842425 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842429 4858 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842433 4858 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842438 4858 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842442 4858 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842445 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842450 4858 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842454 4858 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842458 4858 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842462 4858 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842466 4858 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842470 4858 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842473 4858 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842476 4858 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842480 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842484 4858 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.842491 4858 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842605 4858 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842611 4858 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842615 4858 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842619 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842623 4858 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842626 4858 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842630 4858 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842633 4858 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842637 4858 feature_gate.go:330] unrecognized feature gate: Example Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842641 4858 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842645 4858 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842648 4858 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842652 4858 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842655 4858 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842659 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842663 4858 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842667 4858 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842671 4858 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842676 4858 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842680 4858 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842684 4858 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842688 4858 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842691 4858 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842695 4858 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842699 4858 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842702 4858 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842706 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842709 4858 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842713 4858 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842716 4858 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842720 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842725 4858 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842728 4858 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842732 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842736 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842740 4858 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842743 4858 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842747 4858 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842751 4858 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842755 4858 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842758 4858 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842762 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842766 4858 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842770 4858 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842773 4858 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842777 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842781 4858 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842784 4858 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842788 4858 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842792 4858 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842796 4858 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842800 4858 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842804 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842809 4858 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842813 4858 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842817 4858 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842821 4858 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842824 4858 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842828 4858 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842832 4858 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842835 4858 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842839 4858 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842842 4858 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842846 4858 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842850 4858 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842853 4858 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842857 4858 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842860 4858 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842865 4858 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842869 4858 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.842874 4858 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.842880 4858 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.842987 4858 server.go:940] "Client rotation is on, will bootstrap in background" Mar 20 08:56:59 crc kubenswrapper[4858]: E0320 08:56:59.850034 4858 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.853475 4858 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.853568 4858 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.855197 4858 server.go:997] "Starting client certificate rotation" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.855229 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.855395 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.890620 4858 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 08:56:59 crc kubenswrapper[4858]: E0320 08:56:59.893132 4858 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.894077 4858 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.909501 4858 log.go:25] "Validated CRI v1 runtime API" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.946422 4858 log.go:25] "Validated CRI v1 image API" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.948759 4858 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.953233 4858 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-20-08-51-57-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.953282 4858 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.972701 4858 manager.go:217] Machine: {Timestamp:2026-03-20 08:56:59.971049149 +0000 UTC m=+1.291467356 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:3e03dc76-aa4b-4c6f-a4c0-977607dcbe31 BootID:c2684611-1609-4d43-a887-40f21805a2dc Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:ca:1d:d5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:ca:1d:d5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d8:dc:56 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:72:1c:b6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a4:9f:1a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:49:47:c0 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:d2:bb:17 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:7a:a1:83:94:1e:f3 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:56:e2:19:3b:ee:59 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.972908 4858 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.973040 4858 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.973345 4858 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.973507 4858 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.973542 4858 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.974658 4858 topology_manager.go:138] "Creating topology manager with none policy" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.974676 4858 container_manager_linux.go:303] "Creating device plugin manager" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.975012 4858 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.975030 4858 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.975173 4858 state_mem.go:36] "Initialized new in-memory state store" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.975578 4858 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.979290 4858 kubelet.go:418] "Attempting to sync node with API server" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.979374 4858 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.979392 4858 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.979404 4858 kubelet.go:324] "Adding apiserver pod source" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.979415 4858 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.982683 4858 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.983847 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:56:59 crc kubenswrapper[4858]: W0320 08:56:59.983880 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:56:59 crc kubenswrapper[4858]: E0320 08:56:59.983952 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:56:59 crc kubenswrapper[4858]: E0320 08:56:59.983966 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.984002 4858 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.986262 4858 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987850 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987879 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987886 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987893 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987905 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987912 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987918 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987929 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987939 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987946 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987955 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.987962 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.991594 4858 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 20 08:56:59 crc kubenswrapper[4858]: I0320 08:56:59.992100 4858 server.go:1280] "Started kubelet" Mar 20 08:56:59 crc systemd[1]: Started Kubernetes Kubelet. Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.000827 4858 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.001135 4858 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.001705 4858 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.002559 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.002576 4858 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.002731 4858 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.002839 4858 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.002855 4858 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.003014 4858 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.003373 4858 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.004221 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.004290 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.004875 4858 factory.go:55] Registering systemd factory Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.004954 4858 factory.go:221] Registration of the systemd container factory successfully Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.005134 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="200ms" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.005833 4858 factory.go:153] Registering CRI-O factory Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.005863 4858 factory.go:221] Registration of the crio container factory successfully Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.005990 4858 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.006028 4858 factory.go:103] Registering Raw factory Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.006052 4858 manager.go:1196] Started watching for new ooms in manager Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.009265 4858 server.go:460] "Adding debug handlers to kubelet server" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.011042 4858 manager.go:319] Starting recovery of all containers Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.009476 4858 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.166:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189e80e4013a5165 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,LastTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019481 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019564 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019602 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019633 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019647 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019683 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019698 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019713 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019730 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019813 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019848 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019862 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019876 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019892 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019905 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019920 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019936 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019953 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019966 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.019991 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020005 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020017 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020031 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020046 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020086 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020103 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020121 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020138 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020174 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020225 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020243 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020259 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020278 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020326 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020344 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020359 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020395 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020410 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020444 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020462 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020477 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020492 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020506 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.020520 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023140 4858 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023750 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023769 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023806 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023820 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023834 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023844 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023856 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023870 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023887 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023907 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023931 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023947 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023963 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023976 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.023989 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024002 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024016 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024030 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024045 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024058 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024070 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024079 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024090 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024100 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024110 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024121 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024156 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024170 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024182 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024195 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024207 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024218 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024230 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024241 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024251 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024263 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024275 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024286 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024297 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024306 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024330 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024341 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024351 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024360 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024369 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024377 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024387 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024397 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024406 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024423 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024442 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024457 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024469 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024481 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024492 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024503 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024515 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024529 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024544 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024554 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024572 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024587 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024601 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024615 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024662 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024675 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024689 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024702 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024716 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024728 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024741 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024752 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024764 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024777 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024792 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024802 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024815 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024827 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024838 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024850 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024867 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024879 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024892 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024904 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024918 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024930 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024944 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024955 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024966 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024976 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024985 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.024995 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025004 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025013 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025021 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025030 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025039 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025048 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025058 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025070 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025080 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025090 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025098 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025107 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025116 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025126 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025135 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025144 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025154 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025164 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025174 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025183 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025193 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025202 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025212 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025222 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025232 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025243 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025253 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025262 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025271 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025280 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025290 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025300 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025340 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025350 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025361 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025369 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025379 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025389 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025398 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025408 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025417 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025426 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025435 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025445 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025454 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025463 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025472 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025480 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025489 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025500 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025509 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025518 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025528 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025538 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025547 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025557 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025565 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025574 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025584 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025593 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025601 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025610 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025619 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025627 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025637 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025647 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025655 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025665 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025674 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025683 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025692 4858 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025702 4858 reconstruct.go:97] "Volume reconstruction finished" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.025710 4858 reconciler.go:26] "Reconciler: start to sync state" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.040137 4858 manager.go:324] Recovery completed Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.058012 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.060182 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.060239 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.060252 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.061966 4858 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.061991 4858 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.062025 4858 state_mem.go:36] "Initialized new in-memory state store" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.065684 4858 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.068657 4858 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.068744 4858 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.068784 4858 kubelet.go:2335] "Starting kubelet main sync loop" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.068840 4858 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.069403 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.069474 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.092090 4858 policy_none.go:49] "None policy: Start" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.093277 4858 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.093515 4858 state_mem.go:35] "Initializing new in-memory state store" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.104061 4858 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.157500 4858 manager.go:334] "Starting Device Plugin manager" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.157553 4858 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.157568 4858 server.go:79] "Starting device plugin registration server" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.157997 4858 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.158017 4858 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.158705 4858 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.158802 4858 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.158815 4858 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.166634 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.169114 4858 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.169180 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.169948 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.169978 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.169989 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.170111 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.170353 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.170388 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.170837 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.170917 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.170975 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.170919 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.171091 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.171102 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.171264 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.171370 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.171540 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.172181 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.172207 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.172218 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.172647 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.172681 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.172691 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.172810 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173033 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173127 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173362 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173390 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173403 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173495 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173636 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.173666 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174127 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174152 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174162 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174298 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174338 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174368 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174345 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174380 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174903 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174921 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174928 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174971 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.174988 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.175000 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.205831 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="400ms" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.227211 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.227680 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.227743 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.227873 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.227942 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228693 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228739 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228786 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228820 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228845 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228874 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228923 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228947 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228974 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.228998 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.259779 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.261005 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.261044 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.261055 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.261082 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.261602 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.166:6443: connect: connection refused" node="crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.330793 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.330883 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.330930 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.330960 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.330991 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331005 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331022 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331053 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331100 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331179 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331209 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331211 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331218 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331239 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331351 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331306 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331401 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331424 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331469 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331488 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331518 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331575 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331530 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331552 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331594 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331652 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331535 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331688 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331748 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.331841 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.462040 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.464012 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.464048 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.464057 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.464079 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.464585 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.166:6443: connect: connection refused" node="crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.502388 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.522835 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.534459 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.553126 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-50952a9614e293613e753fa44c3c7410fa08fa638719c191f8b7b59a7087cb66 WatchSource:0}: Error finding container 50952a9614e293613e753fa44c3c7410fa08fa638719c191f8b7b59a7087cb66: Status 404 returned error can't find the container with id 50952a9614e293613e753fa44c3c7410fa08fa638719c191f8b7b59a7087cb66 Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.554593 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-9792691ada9e30e330e94a3358dbf698c698622f5b7da427a2485a5373310b89 WatchSource:0}: Error finding container 9792691ada9e30e330e94a3358dbf698c698622f5b7da427a2485a5373310b89: Status 404 returned error can't find the container with id 9792691ada9e30e330e94a3358dbf698c698622f5b7da427a2485a5373310b89 Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.561161 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.562236 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-5ee3d292feafe47616bea6adf35c13ea746cdfcf80154ddd104fdb4218c09e1a WatchSource:0}: Error finding container 5ee3d292feafe47616bea6adf35c13ea746cdfcf80154ddd104fdb4218c09e1a: Status 404 returned error can't find the container with id 5ee3d292feafe47616bea6adf35c13ea746cdfcf80154ddd104fdb4218c09e1a Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.568866 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.574678 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-4ef7d07d714ec9dfabd30fb2da830e670121d708f0092201a29cb142c1f74eb3 WatchSource:0}: Error finding container 4ef7d07d714ec9dfabd30fb2da830e670121d708f0092201a29cb142c1f74eb3: Status 404 returned error can't find the container with id 4ef7d07d714ec9dfabd30fb2da830e670121d708f0092201a29cb142c1f74eb3 Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.592561 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-ca0573a48a78e211ad97d86902aab41a4f5224552a1b918f15ad34a3df82a6d9 WatchSource:0}: Error finding container ca0573a48a78e211ad97d86902aab41a4f5224552a1b918f15ad34a3df82a6d9: Status 404 returned error can't find the container with id ca0573a48a78e211ad97d86902aab41a4f5224552a1b918f15ad34a3df82a6d9 Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.607820 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="800ms" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.865382 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.869079 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.869128 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.869143 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:00 crc kubenswrapper[4858]: I0320 08:57:00.869172 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.869664 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.166:6443: connect: connection refused" node="crc" Mar 20 08:57:00 crc kubenswrapper[4858]: W0320 08:57:00.963876 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:00 crc kubenswrapper[4858]: E0320 08:57:00.964002 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.003614 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.074543 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5ee3d292feafe47616bea6adf35c13ea746cdfcf80154ddd104fdb4218c09e1a"} Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.075375 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"50952a9614e293613e753fa44c3c7410fa08fa638719c191f8b7b59a7087cb66"} Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.076394 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9792691ada9e30e330e94a3358dbf698c698622f5b7da427a2485a5373310b89"} Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.077204 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ca0573a48a78e211ad97d86902aab41a4f5224552a1b918f15ad34a3df82a6d9"} Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.078543 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4ef7d07d714ec9dfabd30fb2da830e670121d708f0092201a29cb142c1f74eb3"} Mar 20 08:57:01 crc kubenswrapper[4858]: W0320 08:57:01.093088 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:01 crc kubenswrapper[4858]: E0320 08:57:01.093179 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:01 crc kubenswrapper[4858]: E0320 08:57:01.165102 4858 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.166:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189e80e4013a5165 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,LastTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:01 crc kubenswrapper[4858]: W0320 08:57:01.300853 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:01 crc kubenswrapper[4858]: E0320 08:57:01.300998 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:01 crc kubenswrapper[4858]: E0320 08:57:01.409645 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="1.6s" Mar 20 08:57:01 crc kubenswrapper[4858]: W0320 08:57:01.581877 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:01 crc kubenswrapper[4858]: E0320 08:57:01.581969 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.670656 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.672746 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.672814 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.672830 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.672872 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:01 crc kubenswrapper[4858]: E0320 08:57:01.673605 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.166:6443: connect: connection refused" node="crc" Mar 20 08:57:01 crc kubenswrapper[4858]: I0320 08:57:01.967820 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:01 crc kubenswrapper[4858]: E0320 08:57:01.968768 4858 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.004447 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.082224 4858 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e" exitCode=0 Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.082368 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.082429 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.083105 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.083132 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.083142 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.084468 4858 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d" exitCode=0 Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.084515 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.084565 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.086725 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.086782 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.086802 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.091652 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.091683 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.091694 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.091703 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.091735 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.092922 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.092973 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.092989 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.093213 4858 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f45e5e430fe4e7a16b7273395bca040e6028738ab31ad5a558815c648bad9fce" exitCode=0 Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.093277 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f45e5e430fe4e7a16b7273395bca040e6028738ab31ad5a558815c648bad9fce"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.093301 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.094212 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.094245 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.094254 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.094793 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca" exitCode=0 Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.094825 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca"} Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.094924 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.101932 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.101959 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.101968 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.104747 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.105522 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.105538 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.105547 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:02 crc kubenswrapper[4858]: I0320 08:57:02.750903 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.003770 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:03 crc kubenswrapper[4858]: E0320 08:57:03.010510 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="3.2s" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.099736 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.099829 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.102133 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.102183 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.102194 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.105785 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.105866 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.105877 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.105880 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.107123 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.107152 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.107165 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.114914 4858 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1978b9df9626cb0dc533bc542bdb97a17edaebd1c7392df4945120a0193612d1" exitCode=0 Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.115052 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1978b9df9626cb0dc533bc542bdb97a17edaebd1c7392df4945120a0193612d1"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.115391 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.116907 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.116944 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.116958 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.122941 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.123064 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.123916 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.123993 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.124054 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9"} Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.131958 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.132032 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.132059 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.274415 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.275974 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.276013 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.276022 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:03 crc kubenswrapper[4858]: I0320 08:57:03.276048 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:03 crc kubenswrapper[4858]: E0320 08:57:03.276413 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.166:6443: connect: connection refused" node="crc" Mar 20 08:57:03 crc kubenswrapper[4858]: W0320 08:57:03.615402 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:03 crc kubenswrapper[4858]: E0320 08:57:03.615490 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:03 crc kubenswrapper[4858]: W0320 08:57:03.793067 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:03 crc kubenswrapper[4858]: E0320 08:57:03.793232 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:03 crc kubenswrapper[4858]: W0320 08:57:03.823155 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:03 crc kubenswrapper[4858]: E0320 08:57:03.823296 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:03 crc kubenswrapper[4858]: W0320 08:57:03.830356 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:03 crc kubenswrapper[4858]: E0320 08:57:03.830479 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.166:6443: connect: connection refused" logger="UnhandledError" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.003651 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.166:6443: connect: connection refused Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.128100 4858 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d0fbe383a8fc12945197b9377b642389f928e271ca17d147168365c2f6ce67d3" exitCode=0 Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.128158 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d0fbe383a8fc12945197b9377b642389f928e271ca17d147168365c2f6ce67d3"} Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.128342 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.129809 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.129839 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.129849 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.129959 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.132205 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bc29f7d33a13f604b313e219db362ef52e0331a249b5a703e1df896bf039a937" exitCode=255 Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.132307 4858 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.132343 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.132359 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.132374 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"bc29f7d33a13f604b313e219db362ef52e0331a249b5a703e1df896bf039a937"} Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.132364 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133088 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133541 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133568 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133579 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133541 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133609 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133622 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133639 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133677 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.133696 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.134504 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.134567 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.134583 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:04 crc kubenswrapper[4858]: I0320 08:57:04.135350 4858 scope.go:117] "RemoveContainer" containerID="bc29f7d33a13f604b313e219db362ef52e0331a249b5a703e1df896bf039a937" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.116689 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.136844 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.138615 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da"} Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.138843 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.138900 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.140712 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.140751 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.140764 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.146228 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"dc25c6c55a010e6e4a10604213b46a556eb0505da0c8f0fae7e7442bcba02b17"} Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.146281 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"37636277350fa9b8e80370355e2299cb9317b8a2199b4c1143b1bbf48d5b2251"} Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.146296 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"656d0082eafe56e84d2dfd2e4922035744e22a75cced6dc35aa75dd754acb7f7"} Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.146309 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f6e9591e8f680a16befd434e8ec92a4a1b20593bcf35a9f1d3947dc3583056bb"} Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.146327 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.147135 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.147202 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.147222 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.751742 4858 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 08:57:05 crc kubenswrapper[4858]: I0320 08:57:05.751842 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.004895 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.154787 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"04d2e6b195dd8cc7a1453f7230857a5c49f0dcbbd4dcab8241148a535d2d230f"} Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.154938 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.154966 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.155026 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.156690 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.156757 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.156775 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.156839 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.156866 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.156877 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.477085 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.478935 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.479020 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.479055 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.479101 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.526789 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:06 crc kubenswrapper[4858]: I0320 08:57:06.737220 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.157223 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.157346 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.158645 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.158689 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.158705 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.158762 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.158870 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.158897 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.836638 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.836865 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.838402 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.838458 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:07 crc kubenswrapper[4858]: I0320 08:57:07.838484 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.161381 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.161456 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.162956 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.163025 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.163051 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.163213 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.163255 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.163273 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.192308 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.192462 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.193625 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.193700 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.193724 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.736752 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.811634 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:08 crc kubenswrapper[4858]: I0320 08:57:08.817206 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.163050 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.163055 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.163987 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.164023 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.164036 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.164891 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.164962 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:09 crc kubenswrapper[4858]: I0320 08:57:09.164986 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:10 crc kubenswrapper[4858]: I0320 08:57:10.165051 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:10 crc kubenswrapper[4858]: I0320 08:57:10.166140 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:10 crc kubenswrapper[4858]: I0320 08:57:10.166189 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:10 crc kubenswrapper[4858]: I0320 08:57:10.166205 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:10 crc kubenswrapper[4858]: E0320 08:57:10.167073 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:11 crc kubenswrapper[4858]: I0320 08:57:11.084508 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 20 08:57:11 crc kubenswrapper[4858]: I0320 08:57:11.084693 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:11 crc kubenswrapper[4858]: I0320 08:57:11.085777 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:11 crc kubenswrapper[4858]: I0320 08:57:11.085827 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:11 crc kubenswrapper[4858]: I0320 08:57:11.085840 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:14 crc kubenswrapper[4858]: I0320 08:57:14.030944 4858 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 20 08:57:14 crc kubenswrapper[4858]: I0320 08:57:14.031305 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 20 08:57:14 crc kubenswrapper[4858]: W0320 08:57:14.739564 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.739660 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.741251 4858 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e80e4013a5165 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,LastTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:14 crc kubenswrapper[4858]: I0320 08:57:14.742892 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z Mar 20 08:57:14 crc kubenswrapper[4858]: W0320 08:57:14.743236 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.743285 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.744439 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 20 08:57:14 crc kubenswrapper[4858]: W0320 08:57:14.745013 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.745174 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4858]: W0320 08:57:14.745195 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.745396 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.747380 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 08:57:14 crc kubenswrapper[4858]: E0320 08:57:14.747903 4858 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:14Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:14 crc kubenswrapper[4858]: I0320 08:57:14.749533 4858 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 08:57:14 crc kubenswrapper[4858]: I0320 08:57:14.749667 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 20 08:57:14 crc kubenswrapper[4858]: I0320 08:57:14.753407 4858 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 08:57:14 crc kubenswrapper[4858]: I0320 08:57:14.753471 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 20 08:57:15 crc kubenswrapper[4858]: I0320 08:57:15.006678 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:15Z is after 2026-02-23T05:33:13Z Mar 20 08:57:15 crc kubenswrapper[4858]: I0320 08:57:15.752584 4858 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 08:57:15 crc kubenswrapper[4858]: I0320 08:57:15.752682 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.005284 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:16Z is after 2026-02-23T05:33:13Z Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.182213 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.182869 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.184535 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da" exitCode=255 Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.184579 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da"} Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.184623 4858 scope.go:117] "RemoveContainer" containerID="bc29f7d33a13f604b313e219db362ef52e0331a249b5a703e1df896bf039a937" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.184795 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.186402 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.186425 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.186433 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.186879 4858 scope.go:117] "RemoveContainer" containerID="d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da" Mar 20 08:57:16 crc kubenswrapper[4858]: E0320 08:57:16.187026 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.531528 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.767952 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.768101 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.768961 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.768983 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.768990 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:16 crc kubenswrapper[4858]: I0320 08:57:16.783835 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.005646 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:17Z is after 2026-02-23T05:33:13Z Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.188577 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.190504 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.190504 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.191304 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.191356 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.191368 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.191750 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.191792 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.191803 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.191860 4858 scope.go:117] "RemoveContainer" containerID="d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da" Mar 20 08:57:17 crc kubenswrapper[4858]: E0320 08:57:17.192025 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:17 crc kubenswrapper[4858]: I0320 08:57:17.196560 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.006128 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:18Z is after 2026-02-23T05:33:13Z Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.192997 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.194526 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.194594 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.194618 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.195553 4858 scope.go:117] "RemoveContainer" containerID="d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da" Mar 20 08:57:18 crc kubenswrapper[4858]: E0320 08:57:18.195849 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.201447 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.201788 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.205186 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.205269 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:18 crc kubenswrapper[4858]: I0320 08:57:18.205289 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:19 crc kubenswrapper[4858]: I0320 08:57:19.008231 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:19Z is after 2026-02-23T05:33:13Z Mar 20 08:57:20 crc kubenswrapper[4858]: I0320 08:57:20.006029 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:20Z is after 2026-02-23T05:33:13Z Mar 20 08:57:20 crc kubenswrapper[4858]: E0320 08:57:20.167196 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:21 crc kubenswrapper[4858]: I0320 08:57:21.006798 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:21Z is after 2026-02-23T05:33:13Z Mar 20 08:57:21 crc kubenswrapper[4858]: I0320 08:57:21.148430 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:21 crc kubenswrapper[4858]: E0320 08:57:21.149594 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:21Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 08:57:21 crc kubenswrapper[4858]: I0320 08:57:21.150173 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:21 crc kubenswrapper[4858]: I0320 08:57:21.150229 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:21 crc kubenswrapper[4858]: I0320 08:57:21.150247 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:21 crc kubenswrapper[4858]: I0320 08:57:21.150286 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:21 crc kubenswrapper[4858]: E0320 08:57:21.153682 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:21Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 08:57:22 crc kubenswrapper[4858]: I0320 08:57:22.007583 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:22Z is after 2026-02-23T05:33:13Z Mar 20 08:57:22 crc kubenswrapper[4858]: I0320 08:57:22.057438 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:22 crc kubenswrapper[4858]: I0320 08:57:22.057852 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:22 crc kubenswrapper[4858]: I0320 08:57:22.059937 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:22 crc kubenswrapper[4858]: I0320 08:57:22.060045 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:22 crc kubenswrapper[4858]: I0320 08:57:22.060065 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:22 crc kubenswrapper[4858]: I0320 08:57:22.061173 4858 scope.go:117] "RemoveContainer" containerID="d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da" Mar 20 08:57:22 crc kubenswrapper[4858]: E0320 08:57:22.061521 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:23 crc kubenswrapper[4858]: I0320 08:57:23.010443 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:23Z is after 2026-02-23T05:33:13Z Mar 20 08:57:23 crc kubenswrapper[4858]: I0320 08:57:23.052702 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:23 crc kubenswrapper[4858]: E0320 08:57:23.056719 4858 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:23Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:24 crc kubenswrapper[4858]: I0320 08:57:24.007244 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:24Z is after 2026-02-23T05:33:13Z Mar 20 08:57:24 crc kubenswrapper[4858]: W0320 08:57:24.586878 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:24Z is after 2026-02-23T05:33:13Z Mar 20 08:57:24 crc kubenswrapper[4858]: E0320 08:57:24.586958 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:24Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:24 crc kubenswrapper[4858]: E0320 08:57:24.747012 4858 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:24Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e80e4013a5165 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,LastTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.007419 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z Mar 20 08:57:25 crc kubenswrapper[4858]: W0320 08:57:25.397168 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z Mar 20 08:57:25 crc kubenswrapper[4858]: E0320 08:57:25.397270 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.752492 4858 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.752616 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.752721 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.752973 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.754301 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.754407 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.754428 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.754936 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 20 08:57:25 crc kubenswrapper[4858]: I0320 08:57:25.755142 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0" gracePeriod=30 Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.006631 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:26Z is after 2026-02-23T05:33:13Z Mar 20 08:57:26 crc kubenswrapper[4858]: W0320 08:57:26.114697 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:26Z is after 2026-02-23T05:33:13Z Mar 20 08:57:26 crc kubenswrapper[4858]: E0320 08:57:26.114768 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.215282 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.215809 4858 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0" exitCode=255 Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.215841 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0"} Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.215992 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3"} Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.216133 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.217031 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.217124 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:26 crc kubenswrapper[4858]: I0320 08:57:26.217192 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:26 crc kubenswrapper[4858]: W0320 08:57:26.728303 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:26Z is after 2026-02-23T05:33:13Z Mar 20 08:57:26 crc kubenswrapper[4858]: E0320 08:57:26.728800 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:27 crc kubenswrapper[4858]: I0320 08:57:27.005808 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:27Z is after 2026-02-23T05:33:13Z Mar 20 08:57:27 crc kubenswrapper[4858]: I0320 08:57:27.836779 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:27 crc kubenswrapper[4858]: I0320 08:57:27.837245 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:27 crc kubenswrapper[4858]: I0320 08:57:27.838739 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:27 crc kubenswrapper[4858]: I0320 08:57:27.838786 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:27 crc kubenswrapper[4858]: I0320 08:57:27.838803 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:28 crc kubenswrapper[4858]: I0320 08:57:28.006384 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z Mar 20 08:57:28 crc kubenswrapper[4858]: I0320 08:57:28.153851 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:28 crc kubenswrapper[4858]: E0320 08:57:28.154116 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 08:57:28 crc kubenswrapper[4858]: I0320 08:57:28.155127 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:28 crc kubenswrapper[4858]: I0320 08:57:28.155169 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:28 crc kubenswrapper[4858]: I0320 08:57:28.155183 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:28 crc kubenswrapper[4858]: I0320 08:57:28.155209 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:28 crc kubenswrapper[4858]: E0320 08:57:28.159949 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:28Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 08:57:29 crc kubenswrapper[4858]: I0320 08:57:29.005599 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:29Z is after 2026-02-23T05:33:13Z Mar 20 08:57:30 crc kubenswrapper[4858]: I0320 08:57:30.005940 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:30Z is after 2026-02-23T05:33:13Z Mar 20 08:57:30 crc kubenswrapper[4858]: E0320 08:57:30.167302 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:31 crc kubenswrapper[4858]: I0320 08:57:31.006300 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:31Z is after 2026-02-23T05:33:13Z Mar 20 08:57:32 crc kubenswrapper[4858]: I0320 08:57:32.006518 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:32Z is after 2026-02-23T05:33:13Z Mar 20 08:57:32 crc kubenswrapper[4858]: I0320 08:57:32.751608 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:32 crc kubenswrapper[4858]: I0320 08:57:32.752105 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:32 crc kubenswrapper[4858]: I0320 08:57:32.753449 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:32 crc kubenswrapper[4858]: I0320 08:57:32.753507 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:32 crc kubenswrapper[4858]: I0320 08:57:32.753519 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:33 crc kubenswrapper[4858]: I0320 08:57:33.006523 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:33Z is after 2026-02-23T05:33:13Z Mar 20 08:57:34 crc kubenswrapper[4858]: I0320 08:57:34.005969 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:34Z is after 2026-02-23T05:33:13Z Mar 20 08:57:34 crc kubenswrapper[4858]: E0320 08:57:34.751474 4858 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:34Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189e80e4013a5165 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,LastTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.006887 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:35Z is after 2026-02-23T05:33:13Z Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.160155 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:35 crc kubenswrapper[4858]: E0320 08:57:35.160200 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:35Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.161676 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.161715 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.161726 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.161754 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:35 crc kubenswrapper[4858]: E0320 08:57:35.165711 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:35Z is after 2026-02-23T05:33:13Z" node="crc" Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.752495 4858 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 08:57:35 crc kubenswrapper[4858]: I0320 08:57:35.752624 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 08:57:36 crc kubenswrapper[4858]: I0320 08:57:36.009451 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:36Z is after 2026-02-23T05:33:13Z Mar 20 08:57:37 crc kubenswrapper[4858]: I0320 08:57:37.008441 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:37Z is after 2026-02-23T05:33:13Z Mar 20 08:57:37 crc kubenswrapper[4858]: I0320 08:57:37.069837 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:37 crc kubenswrapper[4858]: I0320 08:57:37.071059 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:37 crc kubenswrapper[4858]: I0320 08:57:37.071109 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:37 crc kubenswrapper[4858]: I0320 08:57:37.071122 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:37 crc kubenswrapper[4858]: I0320 08:57:37.071748 4858 scope.go:117] "RemoveContainer" containerID="d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.006224 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:38Z is after 2026-02-23T05:33:13Z Mar 20 08:57:38 crc kubenswrapper[4858]: W0320 08:57:38.229550 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:38Z is after 2026-02-23T05:33:13Z Mar 20 08:57:38 crc kubenswrapper[4858]: E0320 08:57:38.229624 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:38Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.243631 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.244218 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.246699 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="944a33d0ded90b4326b3b55757d865edd09e72b284f82106cc2579922482770d" exitCode=255 Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.246742 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"944a33d0ded90b4326b3b55757d865edd09e72b284f82106cc2579922482770d"} Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.246829 4858 scope.go:117] "RemoveContainer" containerID="d075fb1086edd03807d828d4d033c470887a6bfdac18f9e4c38faf9c4dbed0da" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.246955 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.247781 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.247811 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.247822 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:38 crc kubenswrapper[4858]: I0320 08:57:38.248282 4858 scope.go:117] "RemoveContainer" containerID="944a33d0ded90b4326b3b55757d865edd09e72b284f82106cc2579922482770d" Mar 20 08:57:38 crc kubenswrapper[4858]: E0320 08:57:38.248489 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:39 crc kubenswrapper[4858]: I0320 08:57:39.006578 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:39Z is after 2026-02-23T05:33:13Z Mar 20 08:57:39 crc kubenswrapper[4858]: I0320 08:57:39.250156 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:57:40 crc kubenswrapper[4858]: I0320 08:57:40.005998 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:40Z is after 2026-02-23T05:33:13Z Mar 20 08:57:40 crc kubenswrapper[4858]: E0320 08:57:40.167615 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:40 crc kubenswrapper[4858]: I0320 08:57:40.377635 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:57:40 crc kubenswrapper[4858]: E0320 08:57:40.382531 4858 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:40Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 20 08:57:40 crc kubenswrapper[4858]: E0320 08:57:40.383823 4858 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 20 08:57:41 crc kubenswrapper[4858]: I0320 08:57:41.005957 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.010026 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.056698 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.056932 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.058539 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.058598 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.058617 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.059406 4858 scope.go:117] "RemoveContainer" containerID="944a33d0ded90b4326b3b55757d865edd09e72b284f82106cc2579922482770d" Mar 20 08:57:42 crc kubenswrapper[4858]: E0320 08:57:42.059676 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:42 crc kubenswrapper[4858]: E0320 08:57:42.162242 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.166126 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.167267 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.167405 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.167455 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:42 crc kubenswrapper[4858]: I0320 08:57:42.167504 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:42 crc kubenswrapper[4858]: E0320 08:57:42.171822 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:57:43 crc kubenswrapper[4858]: I0320 08:57:43.010451 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:44 crc kubenswrapper[4858]: I0320 08:57:44.011641 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:44 crc kubenswrapper[4858]: I0320 08:57:44.029953 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:57:44 crc kubenswrapper[4858]: I0320 08:57:44.030243 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:44 crc kubenswrapper[4858]: I0320 08:57:44.031775 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:44 crc kubenswrapper[4858]: I0320 08:57:44.031840 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:44 crc kubenswrapper[4858]: I0320 08:57:44.031860 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:44 crc kubenswrapper[4858]: I0320 08:57:44.032714 4858 scope.go:117] "RemoveContainer" containerID="944a33d0ded90b4326b3b55757d865edd09e72b284f82106cc2579922482770d" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.032991 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:57:44 crc kubenswrapper[4858]: W0320 08:57:44.546684 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.546770 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.758904 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4013a5165 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,LastTimestamp:2026-03-20 08:56:59.992052069 +0000 UTC m=+1.312470286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.764498 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.771544 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.778420 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054b14a5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,LastTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.783180 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e40b40e524 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.160255268 +0000 UTC m=+1.480673465,LastTimestamp:2026-03-20 08:57:00.160255268 +0000 UTC m=+1.480673465,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.790655 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054a810f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.169965278 +0000 UTC m=+1.490383475,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.798391 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054ae768\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.169984568 +0000 UTC m=+1.490402765,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.805085 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054b14a5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054b14a5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,LastTimestamp:2026-03-20 08:57:00.169994429 +0000 UTC m=+1.490412626,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.812272 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054a810f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.170906404 +0000 UTC m=+1.491324591,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.822564 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054ae768\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.170969117 +0000 UTC m=+1.491387314,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.828311 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054b14a5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054b14a5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,LastTimestamp:2026-03-20 08:57:00.171040439 +0000 UTC m=+1.491458636,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.836033 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054a810f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.171081111 +0000 UTC m=+1.491499308,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.843982 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054ae768\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.171098832 +0000 UTC m=+1.491517029,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.852171 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054b14a5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054b14a5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,LastTimestamp:2026-03-20 08:57:00.171106662 +0000 UTC m=+1.491524859,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.858590 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054a810f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.172200314 +0000 UTC m=+1.492618511,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.865928 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054ae768\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.172213915 +0000 UTC m=+1.492632112,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.873293 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054b14a5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054b14a5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,LastTimestamp:2026-03-20 08:57:00.172224425 +0000 UTC m=+1.492642622,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.880103 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054a810f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.172667442 +0000 UTC m=+1.493085639,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.886873 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054ae768\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.172687993 +0000 UTC m=+1.493106190,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.893918 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054b14a5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054b14a5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,LastTimestamp:2026-03-20 08:57:00.172696294 +0000 UTC m=+1.493114491,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.900875 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054a810f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.173377021 +0000 UTC m=+1.493795218,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.907661 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054ae768\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.173398752 +0000 UTC m=+1.493816949,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.913256 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054b14a5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054b14a5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060259493 +0000 UTC m=+1.380677710,LastTimestamp:2026-03-20 08:57:00.173408572 +0000 UTC m=+1.493826769,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.920082 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054a810f\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054a810f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060221711 +0000 UTC m=+1.380639918,LastTimestamp:2026-03-20 08:57:00.17414433 +0000 UTC m=+1.494562527,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.927967 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189e80e4054ae768\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189e80e4054ae768 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.060247912 +0000 UTC m=+1.380666129,LastTimestamp:2026-03-20 08:57:00.174158591 +0000 UTC m=+1.494576788,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.932009 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e42325cf79 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.561133433 +0000 UTC m=+1.881551680,LastTimestamp:2026-03-20 08:57:00.561133433 +0000 UTC m=+1.881551680,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.938151 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e42326da34 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.561201716 +0000 UTC m=+1.881619933,LastTimestamp:2026-03-20 08:57:00.561201716 +0000 UTC m=+1.881619933,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.943621 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4236a8fc6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.56563911 +0000 UTC m=+1.886057307,LastTimestamp:2026-03-20 08:57:00.56563911 +0000 UTC m=+1.886057307,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.948090 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e424362763 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.578981731 +0000 UTC m=+1.899399938,LastTimestamp:2026-03-20 08:57:00.578981731 +0000 UTC m=+1.899399938,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.951540 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e4254f9d68 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:00.59742756 +0000 UTC m=+1.917845767,LastTimestamp:2026-03-20 08:57:00.59742756 +0000 UTC m=+1.917845767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.955989 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e44585d92a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.137852714 +0000 UTC m=+2.458270911,LastTimestamp:2026-03-20 08:57:01.137852714 +0000 UTC m=+2.458270911,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.960512 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e44586be3b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.137911355 +0000 UTC m=+2.458329552,LastTimestamp:2026-03-20 08:57:01.137911355 +0000 UTC m=+2.458329552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.965711 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e44587a36b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.137970027 +0000 UTC m=+2.458388224,LastTimestamp:2026-03-20 08:57:01.137970027 +0000 UTC m=+2.458388224,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.969511 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4458a5ac0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.138148032 +0000 UTC m=+2.458566229,LastTimestamp:2026-03-20 08:57:01.138148032 +0000 UTC m=+2.458566229,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.973311 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e4458e0cd4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.138390228 +0000 UTC m=+2.458808425,LastTimestamp:2026-03-20 08:57:01.138390228 +0000 UTC m=+2.458808425,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.977175 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e4463ae82a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.14971857 +0000 UTC m=+2.470136767,LastTimestamp:2026-03-20 08:57:01.14971857 +0000 UTC m=+2.470136767,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.980759 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e446865798 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.154662296 +0000 UTC m=+2.475080493,LastTimestamp:2026-03-20 08:57:01.154662296 +0000 UTC m=+2.475080493,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.984546 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4469f38b7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.156292791 +0000 UTC m=+2.476710988,LastTimestamp:2026-03-20 08:57:01.156292791 +0000 UTC m=+2.476710988,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.988462 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e446a3a91d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.156583709 +0000 UTC m=+2.477001906,LastTimestamp:2026-03-20 08:57:01.156583709 +0000 UTC m=+2.477001906,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.992683 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e4474ed86e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.167802478 +0000 UTC m=+2.488220675,LastTimestamp:2026-03-20 08:57:01.167802478 +0000 UTC m=+2.488220675,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:44 crc kubenswrapper[4858]: E0320 08:57:44.996710 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e447b7be22 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.174677026 +0000 UTC m=+2.495095213,LastTimestamp:2026-03-20 08:57:01.174677026 +0000 UTC m=+2.495095213,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.000845 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e45af144b9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.497214137 +0000 UTC m=+2.817632374,LastTimestamp:2026-03-20 08:57:01.497214137 +0000 UTC m=+2.817632374,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.004603 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e45bd8fde3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.512400355 +0000 UTC m=+2.832818552,LastTimestamp:2026-03-20 08:57:01.512400355 +0000 UTC m=+2.832818552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.008914 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e45be54055 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.513203797 +0000 UTC m=+2.833622024,LastTimestamp:2026-03-20 08:57:01.513203797 +0000 UTC m=+2.833622024,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: I0320 08:57:45.009016 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.010616 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e468af7db1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.727784369 +0000 UTC m=+3.048202566,LastTimestamp:2026-03-20 08:57:01.727784369 +0000 UTC m=+3.048202566,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.014471 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e4699e634f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.743440719 +0000 UTC m=+3.063858916,LastTimestamp:2026-03-20 08:57:01.743440719 +0000 UTC m=+3.063858916,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.018983 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e469b1f4d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.744723155 +0000 UTC m=+3.065141362,LastTimestamp:2026-03-20 08:57:01.744723155 +0000 UTC m=+3.065141362,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.023823 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e474c16a64 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.930285668 +0000 UTC m=+3.250703905,LastTimestamp:2026-03-20 08:57:01.930285668 +0000 UTC m=+3.250703905,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.027986 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e4756118a0 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.940750496 +0000 UTC m=+3.261168733,LastTimestamp:2026-03-20 08:57:01.940750496 +0000 UTC m=+3.261168733,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.031877 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e47df8e94d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.084917581 +0000 UTC m=+3.405335778,LastTimestamp:2026-03-20 08:57:02.084917581 +0000 UTC m=+3.405335778,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.036095 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e47e2aa630 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.0881772 +0000 UTC m=+3.408595437,LastTimestamp:2026-03-20 08:57:02.0881772 +0000 UTC m=+3.408595437,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.040663 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e47f1ad280 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.103917184 +0000 UTC m=+3.424335371,LastTimestamp:2026-03-20 08:57:02.103917184 +0000 UTC m=+3.424335371,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.044876 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e47f25e5d9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.104643033 +0000 UTC m=+3.425061230,LastTimestamp:2026-03-20 08:57:02.104643033 +0000 UTC m=+3.425061230,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.049114 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e48ca16fff openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.330843135 +0000 UTC m=+3.651261332,LastTimestamp:2026-03-20 08:57:02.330843135 +0000 UTC m=+3.651261332,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.053004 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e48caa5be3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.331427811 +0000 UTC m=+3.651846008,LastTimestamp:2026-03-20 08:57:02.331427811 +0000 UTC m=+3.651846008,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.056390 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e48cb6c673 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.332241523 +0000 UTC m=+3.652659720,LastTimestamp:2026-03-20 08:57:02.332241523 +0000 UTC m=+3.652659720,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.060123 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e48cbe720f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.332744207 +0000 UTC m=+3.653162404,LastTimestamp:2026-03-20 08:57:02.332744207 +0000 UTC m=+3.653162404,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.064816 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e48d72784a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.344542282 +0000 UTC m=+3.664960479,LastTimestamp:2026-03-20 08:57:02.344542282 +0000 UTC m=+3.664960479,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.070149 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e48d8516ab openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.345762475 +0000 UTC m=+3.666180672,LastTimestamp:2026-03-20 08:57:02.345762475 +0000 UTC m=+3.666180672,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.074389 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189e80e48de0fbba openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.35178489 +0000 UTC m=+3.672203107,LastTimestamp:2026-03-20 08:57:02.35178489 +0000 UTC m=+3.672203107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.078492 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e48de399ab openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.351956395 +0000 UTC m=+3.672374592,LastTimestamp:2026-03-20 08:57:02.351956395 +0000 UTC m=+3.672374592,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.082062 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e48df8150e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.353298702 +0000 UTC m=+3.673716899,LastTimestamp:2026-03-20 08:57:02.353298702 +0000 UTC m=+3.673716899,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.086020 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e48e2fbd74 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.356946292 +0000 UTC m=+3.677364489,LastTimestamp:2026-03-20 08:57:02.356946292 +0000 UTC m=+3.677364489,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.089936 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e499c63878 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.55135756 +0000 UTC m=+3.871775777,LastTimestamp:2026-03-20 08:57:02.55135756 +0000 UTC m=+3.871775777,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.093986 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e499c85440 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.551495744 +0000 UTC m=+3.871913941,LastTimestamp:2026-03-20 08:57:02.551495744 +0000 UTC m=+3.871913941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.099579 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e49a7b81c1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.563238337 +0000 UTC m=+3.883656544,LastTimestamp:2026-03-20 08:57:02.563238337 +0000 UTC m=+3.883656544,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.104057 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e49a8abeb8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.564236984 +0000 UTC m=+3.884655201,LastTimestamp:2026-03-20 08:57:02.564236984 +0000 UTC m=+3.884655201,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.108368 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e49aa883fc openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.566188028 +0000 UTC m=+3.886606235,LastTimestamp:2026-03-20 08:57:02.566188028 +0000 UTC m=+3.886606235,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.111974 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e49ab347a5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.566893477 +0000 UTC m=+3.887311684,LastTimestamp:2026-03-20 08:57:02.566893477 +0000 UTC m=+3.887311684,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.116192 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4a775caa1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.780967585 +0000 UTC m=+4.101385792,LastTimestamp:2026-03-20 08:57:02.780967585 +0000 UTC m=+4.101385792,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.122062 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e4a81e0110 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.791991568 +0000 UTC m=+4.112409775,LastTimestamp:2026-03-20 08:57:02.791991568 +0000 UTC m=+4.112409775,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.128231 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4a891d2f0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.799581936 +0000 UTC m=+4.120000143,LastTimestamp:2026-03-20 08:57:02.799581936 +0000 UTC m=+4.120000143,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.132702 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4a8a552ea openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.800859882 +0000 UTC m=+4.121278099,LastTimestamp:2026-03-20 08:57:02.800859882 +0000 UTC m=+4.121278099,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.136664 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189e80e4a8d96dc2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.804274626 +0000 UTC m=+4.124692853,LastTimestamp:2026-03-20 08:57:02.804274626 +0000 UTC m=+4.124692853,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.140551 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4b397f0a5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.984532133 +0000 UTC m=+4.304950330,LastTimestamp:2026-03-20 08:57:02.984532133 +0000 UTC m=+4.304950330,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.144961 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4b468fe54 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.99823266 +0000 UTC m=+4.318650857,LastTimestamp:2026-03-20 08:57:02.99823266 +0000 UTC m=+4.318650857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.149430 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4b4772ac9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.999161545 +0000 UTC m=+4.319579742,LastTimestamp:2026-03-20 08:57:02.999161545 +0000 UTC m=+4.319579742,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.155096 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e4bb981539 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:03.118759225 +0000 UTC m=+4.439177432,LastTimestamp:2026-03-20 08:57:03.118759225 +0000 UTC m=+4.439177432,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.159735 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4bf30d767 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:03.179102055 +0000 UTC m=+4.499520262,LastTimestamp:2026-03-20 08:57:03.179102055 +0000 UTC m=+4.499520262,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.163632 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4bff7f4b7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:03.192151223 +0000 UTC m=+4.512569430,LastTimestamp:2026-03-20 08:57:03.192151223 +0000 UTC m=+4.512569430,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.168162 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e4d0288938 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:03.463770424 +0000 UTC m=+4.784188621,LastTimestamp:2026-03-20 08:57:03.463770424 +0000 UTC m=+4.784188621,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.172696 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e4d0f10a5d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:03.476910685 +0000 UTC m=+4.797328882,LastTimestamp:2026-03-20 08:57:03.476910685 +0000 UTC m=+4.797328882,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.178114 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e4f7f1a471 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.131261553 +0000 UTC m=+5.451679750,LastTimestamp:2026-03-20 08:57:04.131261553 +0000 UTC m=+5.451679750,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.184614 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e80e4b4772ac9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4b4772ac9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:02.999161545 +0000 UTC m=+4.319579742,LastTimestamp:2026-03-20 08:57:04.138615584 +0000 UTC m=+5.459033781,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.189353 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e80e4bf30d767\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4bf30d767 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:03.179102055 +0000 UTC m=+4.499520262,LastTimestamp:2026-03-20 08:57:04.353256458 +0000 UTC m=+5.673674655,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.193947 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e5052e90a7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.353357991 +0000 UTC m=+5.673776188,LastTimestamp:2026-03-20 08:57:04.353357991 +0000 UTC m=+5.673776188,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.199422 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e505ae0110 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.36170984 +0000 UTC m=+5.682128037,LastTimestamp:2026-03-20 08:57:04.36170984 +0000 UTC m=+5.682128037,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.204923 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e505be5f23 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.362782499 +0000 UTC m=+5.683200696,LastTimestamp:2026-03-20 08:57:04.362782499 +0000 UTC m=+5.683200696,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.210490 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189e80e4bff7f4b7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e4bff7f4b7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:03.192151223 +0000 UTC m=+4.512569430,LastTimestamp:2026-03-20 08:57:04.363832489 +0000 UTC m=+5.684250686,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.215178 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e510ae2b5e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.546270046 +0000 UTC m=+5.866688243,LastTimestamp:2026-03-20 08:57:04.546270046 +0000 UTC m=+5.866688243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.219952 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e511894f63 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.560631651 +0000 UTC m=+5.881049848,LastTimestamp:2026-03-20 08:57:04.560631651 +0000 UTC m=+5.881049848,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.225153 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e5119ae199 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.561783193 +0000 UTC m=+5.882201390,LastTimestamp:2026-03-20 08:57:04.561783193 +0000 UTC m=+5.882201390,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.229489 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e51c1ee36f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.738206575 +0000 UTC m=+6.058624782,LastTimestamp:2026-03-20 08:57:04.738206575 +0000 UTC m=+6.058624782,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.235930 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e51ce00845 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.750864453 +0000 UTC m=+6.071282650,LastTimestamp:2026-03-20 08:57:04.750864453 +0000 UTC m=+6.071282650,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.238809 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e51ceeeccc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.75184046 +0000 UTC m=+6.072258657,LastTimestamp:2026-03-20 08:57:04.75184046 +0000 UTC m=+6.072258657,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.242751 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e5276af266 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.927740518 +0000 UTC m=+6.248158735,LastTimestamp:2026-03-20 08:57:04.927740518 +0000 UTC m=+6.248158735,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.248535 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e52862d627 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.943986215 +0000 UTC m=+6.264404412,LastTimestamp:2026-03-20 08:57:04.943986215 +0000 UTC m=+6.264404412,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.253161 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e52877d380 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:04.945361792 +0000 UTC m=+6.265779989,LastTimestamp:2026-03-20 08:57:04.945361792 +0000 UTC m=+6.265779989,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.258520 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e5345bcb45 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.144851269 +0000 UTC m=+6.465269476,LastTimestamp:2026-03-20 08:57:05.144851269 +0000 UTC m=+6.465269476,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.262969 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189e80e535494be3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.160416227 +0000 UTC m=+6.480834434,LastTimestamp:2026-03-20 08:57:05.160416227 +0000 UTC m=+6.480834434,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.268371 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:45 crc kubenswrapper[4858]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80e558895dbf openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 08:57:45 crc kubenswrapper[4858]: body: Mar 20 08:57:45 crc kubenswrapper[4858]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751817663 +0000 UTC m=+7.072235900,LastTimestamp:2026-03-20 08:57:05.751817663 +0000 UTC m=+7.072235900,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:45 crc kubenswrapper[4858]: > Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.273170 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e5588a83ef openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751892975 +0000 UTC m=+7.072311202,LastTimestamp:2026-03-20 08:57:05.751892975 +0000 UTC m=+7.072311202,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.283894 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 08:57:45 crc kubenswrapper[4858]: &Event{ObjectMeta:{kube-apiserver-crc.189e80e7460803ac openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 20 08:57:45 crc kubenswrapper[4858]: body: Mar 20 08:57:45 crc kubenswrapper[4858]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.031285164 +0000 UTC m=+15.351703361,LastTimestamp:2026-03-20 08:57:14.031285164 +0000 UTC m=+15.351703361,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:45 crc kubenswrapper[4858]: > Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.288167 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e746091f46 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.031357766 +0000 UTC m=+15.351775983,LastTimestamp:2026-03-20 08:57:14.031357766 +0000 UTC m=+15.351775983,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.294217 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 20 08:57:45 crc kubenswrapper[4858]: &Event{ObjectMeta:{kube-apiserver-crc.189e80e770d94a61 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 20 08:57:45 crc kubenswrapper[4858]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 20 08:57:45 crc kubenswrapper[4858]: Mar 20 08:57:45 crc kubenswrapper[4858]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.749643361 +0000 UTC m=+16.070061568,LastTimestamp:2026-03-20 08:57:14.749643361 +0000 UTC m=+16.070061568,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:45 crc kubenswrapper[4858]: > Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.300243 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189e80e770db20c8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:14.749763784 +0000 UTC m=+16.070182001,LastTimestamp:2026-03-20 08:57:14.749763784 +0000 UTC m=+16.070182001,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.304715 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e558895dbf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:45 crc kubenswrapper[4858]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80e558895dbf openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 08:57:45 crc kubenswrapper[4858]: body: Mar 20 08:57:45 crc kubenswrapper[4858]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751817663 +0000 UTC m=+7.072235900,LastTimestamp:2026-03-20 08:57:15.752659537 +0000 UTC m=+17.073077754,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:45 crc kubenswrapper[4858]: > Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.308602 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e5588a83ef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e5588a83ef openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751892975 +0000 UTC m=+7.072311202,LastTimestamp:2026-03-20 08:57:15.752707948 +0000 UTC m=+17.073126145,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.315108 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e558895dbf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:45 crc kubenswrapper[4858]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80e558895dbf openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 08:57:45 crc kubenswrapper[4858]: body: Mar 20 08:57:45 crc kubenswrapper[4858]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751817663 +0000 UTC m=+7.072235900,LastTimestamp:2026-03-20 08:57:25.752577719 +0000 UTC m=+27.072995956,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:45 crc kubenswrapper[4858]: > Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.318794 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e5588a83ef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e5588a83ef openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751892975 +0000 UTC m=+7.072311202,LastTimestamp:2026-03-20 08:57:25.752651481 +0000 UTC m=+27.073069708,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.322892 4858 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80ea00d380ab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:25.755117739 +0000 UTC m=+27.075535976,LastTimestamp:2026-03-20 08:57:25.755117739 +0000 UTC m=+27.075535976,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.326921 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e446a3a91d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e446a3a91d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.156583709 +0000 UTC m=+2.477001906,LastTimestamp:2026-03-20 08:57:25.870989615 +0000 UTC m=+27.191407842,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.330480 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e45af144b9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e45af144b9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.497214137 +0000 UTC m=+2.817632374,LastTimestamp:2026-03-20 08:57:26.076022654 +0000 UTC m=+27.396440891,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.334190 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e45bd8fde3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e45bd8fde3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:01.512400355 +0000 UTC m=+2.832818552,LastTimestamp:2026-03-20 08:57:26.088809036 +0000 UTC m=+27.409227233,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.338209 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e558895dbf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:45 crc kubenswrapper[4858]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80e558895dbf openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 08:57:45 crc kubenswrapper[4858]: body: Mar 20 08:57:45 crc kubenswrapper[4858]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751817663 +0000 UTC m=+7.072235900,LastTimestamp:2026-03-20 08:57:35.752589361 +0000 UTC m=+37.073007608,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:45 crc kubenswrapper[4858]: > Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.341484 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e5588a83ef\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189e80e5588a83ef openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751892975 +0000 UTC m=+7.072311202,LastTimestamp:2026-03-20 08:57:35.752674213 +0000 UTC m=+37.073092480,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 08:57:45 crc kubenswrapper[4858]: I0320 08:57:45.752147 4858 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 08:57:45 crc kubenswrapper[4858]: I0320 08:57:45.752288 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 08:57:45 crc kubenswrapper[4858]: E0320 08:57:45.759103 4858 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189e80e558895dbf\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 20 08:57:45 crc kubenswrapper[4858]: &Event{ObjectMeta:{kube-controller-manager-crc.189e80e558895dbf openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 20 08:57:45 crc kubenswrapper[4858]: body: Mar 20 08:57:45 crc kubenswrapper[4858]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 08:57:05.751817663 +0000 UTC m=+7.072235900,LastTimestamp:2026-03-20 08:57:45.752241333 +0000 UTC m=+47.072659590,Count:5,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 20 08:57:45 crc kubenswrapper[4858]: > Mar 20 08:57:46 crc kubenswrapper[4858]: I0320 08:57:46.008055 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:47 crc kubenswrapper[4858]: I0320 08:57:47.006810 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:48 crc kubenswrapper[4858]: I0320 08:57:48.008516 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:48 crc kubenswrapper[4858]: W0320 08:57:48.899821 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 20 08:57:48 crc kubenswrapper[4858]: E0320 08:57:48.900450 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:48 crc kubenswrapper[4858]: W0320 08:57:48.981136 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 20 08:57:48 crc kubenswrapper[4858]: E0320 08:57:48.981226 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 20 08:57:49 crc kubenswrapper[4858]: I0320 08:57:49.010660 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:49 crc kubenswrapper[4858]: E0320 08:57:49.170937 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:57:49 crc kubenswrapper[4858]: I0320 08:57:49.172994 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:49 crc kubenswrapper[4858]: I0320 08:57:49.174887 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:49 crc kubenswrapper[4858]: I0320 08:57:49.174950 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:49 crc kubenswrapper[4858]: I0320 08:57:49.174973 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:49 crc kubenswrapper[4858]: I0320 08:57:49.175029 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:49 crc kubenswrapper[4858]: E0320 08:57:49.180694 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:57:50 crc kubenswrapper[4858]: I0320 08:57:50.008619 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:50 crc kubenswrapper[4858]: E0320 08:57:50.167765 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:57:51 crc kubenswrapper[4858]: I0320 08:57:51.007279 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:52 crc kubenswrapper[4858]: I0320 08:57:52.007609 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:52 crc kubenswrapper[4858]: I0320 08:57:52.755304 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:52 crc kubenswrapper[4858]: I0320 08:57:52.755587 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:52 crc kubenswrapper[4858]: I0320 08:57:52.757137 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:52 crc kubenswrapper[4858]: I0320 08:57:52.757185 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:52 crc kubenswrapper[4858]: I0320 08:57:52.757200 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:52 crc kubenswrapper[4858]: I0320 08:57:52.759265 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 08:57:53 crc kubenswrapper[4858]: I0320 08:57:53.007802 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:53 crc kubenswrapper[4858]: I0320 08:57:53.289771 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:53 crc kubenswrapper[4858]: I0320 08:57:53.291253 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:53 crc kubenswrapper[4858]: I0320 08:57:53.291397 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:53 crc kubenswrapper[4858]: I0320 08:57:53.291488 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:54 crc kubenswrapper[4858]: I0320 08:57:54.007791 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:55 crc kubenswrapper[4858]: I0320 08:57:55.008370 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:55 crc kubenswrapper[4858]: I0320 08:57:55.122241 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 20 08:57:55 crc kubenswrapper[4858]: I0320 08:57:55.122396 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:55 crc kubenswrapper[4858]: I0320 08:57:55.123267 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:55 crc kubenswrapper[4858]: I0320 08:57:55.123304 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:55 crc kubenswrapper[4858]: I0320 08:57:55.123333 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:56 crc kubenswrapper[4858]: I0320 08:57:56.007197 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:56 crc kubenswrapper[4858]: E0320 08:57:56.176152 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:57:56 crc kubenswrapper[4858]: I0320 08:57:56.181188 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:56 crc kubenswrapper[4858]: I0320 08:57:56.182914 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:56 crc kubenswrapper[4858]: I0320 08:57:56.182964 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:56 crc kubenswrapper[4858]: I0320 08:57:56.182974 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:56 crc kubenswrapper[4858]: I0320 08:57:56.182998 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:57:56 crc kubenswrapper[4858]: E0320 08:57:56.188291 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:57:57 crc kubenswrapper[4858]: I0320 08:57:57.006799 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:58 crc kubenswrapper[4858]: I0320 08:57:58.006707 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.006603 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.069197 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.070215 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.070261 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.070270 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.070700 4858 scope.go:117] "RemoveContainer" containerID="944a33d0ded90b4326b3b55757d865edd09e72b284f82106cc2579922482770d" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.306021 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.307804 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208"} Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.307939 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.308691 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.308735 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:57:59 crc kubenswrapper[4858]: I0320 08:57:59.308748 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.006572 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:00 crc kubenswrapper[4858]: E0320 08:58:00.169005 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.311756 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.312178 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.313818 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" exitCode=255 Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.313862 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208"} Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.313949 4858 scope.go:117] "RemoveContainer" containerID="944a33d0ded90b4326b3b55757d865edd09e72b284f82106cc2579922482770d" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.314046 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.314741 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.314763 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.314772 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:00 crc kubenswrapper[4858]: I0320 08:58:00.315252 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 08:58:00 crc kubenswrapper[4858]: E0320 08:58:00.315400 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:01 crc kubenswrapper[4858]: I0320 08:58:01.007378 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:01 crc kubenswrapper[4858]: I0320 08:58:01.317101 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 08:58:02 crc kubenswrapper[4858]: I0320 08:58:02.010245 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:02 crc kubenswrapper[4858]: I0320 08:58:02.057060 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:58:02 crc kubenswrapper[4858]: I0320 08:58:02.057230 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:02 crc kubenswrapper[4858]: I0320 08:58:02.058428 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:02 crc kubenswrapper[4858]: I0320 08:58:02.058449 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:02 crc kubenswrapper[4858]: I0320 08:58:02.058457 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:02 crc kubenswrapper[4858]: I0320 08:58:02.058955 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 08:58:02 crc kubenswrapper[4858]: E0320 08:58:02.059112 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:03 crc kubenswrapper[4858]: I0320 08:58:03.007741 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:03 crc kubenswrapper[4858]: E0320 08:58:03.181796 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:58:03 crc kubenswrapper[4858]: I0320 08:58:03.188816 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:03 crc kubenswrapper[4858]: I0320 08:58:03.190160 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:03 crc kubenswrapper[4858]: I0320 08:58:03.190240 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:03 crc kubenswrapper[4858]: I0320 08:58:03.190257 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:03 crc kubenswrapper[4858]: I0320 08:58:03.190304 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:03 crc kubenswrapper[4858]: E0320 08:58:03.194642 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:58:04 crc kubenswrapper[4858]: I0320 08:58:04.009360 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:04 crc kubenswrapper[4858]: I0320 08:58:04.030257 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:58:04 crc kubenswrapper[4858]: I0320 08:58:04.030637 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:04 crc kubenswrapper[4858]: I0320 08:58:04.032747 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:04 crc kubenswrapper[4858]: I0320 08:58:04.032992 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:04 crc kubenswrapper[4858]: I0320 08:58:04.033014 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:04 crc kubenswrapper[4858]: I0320 08:58:04.033999 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 08:58:04 crc kubenswrapper[4858]: E0320 08:58:04.034297 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:05 crc kubenswrapper[4858]: I0320 08:58:05.007411 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:06 crc kubenswrapper[4858]: I0320 08:58:06.008886 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:07 crc kubenswrapper[4858]: I0320 08:58:07.008448 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:08 crc kubenswrapper[4858]: I0320 08:58:08.007952 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:09 crc kubenswrapper[4858]: I0320 08:58:09.008280 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:10 crc kubenswrapper[4858]: I0320 08:58:10.009135 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:10 crc kubenswrapper[4858]: E0320 08:58:10.169954 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:58:10 crc kubenswrapper[4858]: E0320 08:58:10.186886 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:58:10 crc kubenswrapper[4858]: I0320 08:58:10.194975 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:10 crc kubenswrapper[4858]: I0320 08:58:10.196132 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:10 crc kubenswrapper[4858]: I0320 08:58:10.196174 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:10 crc kubenswrapper[4858]: I0320 08:58:10.196184 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:10 crc kubenswrapper[4858]: I0320 08:58:10.196207 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:10 crc kubenswrapper[4858]: E0320 08:58:10.199988 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:58:11 crc kubenswrapper[4858]: I0320 08:58:11.009720 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:12 crc kubenswrapper[4858]: I0320 08:58:12.010851 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:12 crc kubenswrapper[4858]: I0320 08:58:12.385679 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 20 08:58:12 crc kubenswrapper[4858]: I0320 08:58:12.399498 4858 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 20 08:58:13 crc kubenswrapper[4858]: I0320 08:58:13.010838 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:14 crc kubenswrapper[4858]: I0320 08:58:14.008450 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:15 crc kubenswrapper[4858]: I0320 08:58:15.008303 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:16 crc kubenswrapper[4858]: I0320 08:58:16.007901 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:17 crc kubenswrapper[4858]: I0320 08:58:17.011349 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:17 crc kubenswrapper[4858]: E0320 08:58:17.192221 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 20 08:58:17 crc kubenswrapper[4858]: I0320 08:58:17.200117 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:17 crc kubenswrapper[4858]: I0320 08:58:17.202937 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:17 crc kubenswrapper[4858]: I0320 08:58:17.202969 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:17 crc kubenswrapper[4858]: I0320 08:58:17.202981 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:17 crc kubenswrapper[4858]: I0320 08:58:17.203005 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:17 crc kubenswrapper[4858]: E0320 08:58:17.210169 4858 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 20 08:58:18 crc kubenswrapper[4858]: I0320 08:58:18.009297 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:18 crc kubenswrapper[4858]: I0320 08:58:18.069674 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:18 crc kubenswrapper[4858]: I0320 08:58:18.071460 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:18 crc kubenswrapper[4858]: I0320 08:58:18.071541 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:18 crc kubenswrapper[4858]: I0320 08:58:18.071555 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:18 crc kubenswrapper[4858]: I0320 08:58:18.072237 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 08:58:18 crc kubenswrapper[4858]: E0320 08:58:18.072455 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:19 crc kubenswrapper[4858]: I0320 08:58:19.009029 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:20 crc kubenswrapper[4858]: I0320 08:58:20.006960 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:20 crc kubenswrapper[4858]: E0320 08:58:20.170927 4858 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 20 08:58:20 crc kubenswrapper[4858]: W0320 08:58:20.325834 4858 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 20 08:58:20 crc kubenswrapper[4858]: E0320 08:58:20.325917 4858 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 20 08:58:21 crc kubenswrapper[4858]: I0320 08:58:21.008242 4858 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 20 08:58:21 crc kubenswrapper[4858]: I0320 08:58:21.983895 4858 csr.go:261] certificate signing request csr-mmv22 is approved, waiting to be issued Mar 20 08:58:21 crc kubenswrapper[4858]: I0320 08:58:21.995224 4858 csr.go:257] certificate signing request csr-mmv22 is issued Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.014199 4858 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.069444 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.070988 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.071059 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.071076 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.165798 4858 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.856114 4858 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 20 08:58:22 crc kubenswrapper[4858]: W0320 08:58:22.856303 4858 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.996197 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-28 13:17:59.060522495 +0000 UTC Mar 20 08:58:22 crc kubenswrapper[4858]: I0320 08:58:22.996240 4858 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6076h19m36.064285053s for next certificate rotation Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.035110 4858 apiserver.go:52] "Watching apiserver" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.045403 4858 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.045775 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.046289 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.046344 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.046439 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.046290 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.046651 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.046812 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.046823 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.046915 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.047298 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.049664 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.050010 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.050368 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.050542 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.051271 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.051380 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.051551 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.051731 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.051754 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.086842 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.101140 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.104514 4858 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.111469 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.120364 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.122865 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.122908 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.122956 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.122988 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123016 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123041 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123122 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123156 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123187 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123212 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123234 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123257 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123279 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123302 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123345 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123370 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123393 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123421 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123416 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123450 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123494 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123531 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123540 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123562 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123667 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123685 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123707 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123725 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123740 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123758 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123776 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123798 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123817 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123837 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123855 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123875 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123895 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123916 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123935 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123957 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123978 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123974 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.123998 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124027 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124046 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124061 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124077 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124097 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124092 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124114 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124133 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124129 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124150 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124209 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124227 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124245 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124263 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124280 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124295 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124311 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124340 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124357 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124376 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124395 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124464 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124485 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124508 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124533 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124595 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124616 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124613 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124606 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124606 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124635 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124781 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124850 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124850 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124975 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.124978 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125007 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125038 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125067 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125091 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125115 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125141 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125166 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125191 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125214 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125242 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125266 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125291 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125331 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125356 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125378 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125409 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125437 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125462 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125474 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125487 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125543 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125575 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125603 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125685 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125693 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125723 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125743 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125876 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125935 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.125975 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126013 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126028 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126088 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126156 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126186 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126259 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126365 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126403 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126405 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126456 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.126620 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:23.626571476 +0000 UTC m=+84.946989883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127717 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127742 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127778 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127819 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127847 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127882 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127908 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127939 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127969 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128002 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128037 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128065 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128100 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128130 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128160 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128225 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129073 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129115 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129149 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129183 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129210 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129281 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129326 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129353 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129378 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129485 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129695 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129722 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129748 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129775 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129799 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129824 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129850 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129878 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129904 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129934 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129961 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129984 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130009 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130039 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130063 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130087 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130112 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130140 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130163 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130189 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130212 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130238 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130265 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130294 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130342 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130375 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130402 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130425 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130451 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130475 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130544 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130576 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130603 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130626 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130652 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130680 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130704 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130737 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130764 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130791 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130819 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130849 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130878 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130902 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130931 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130957 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.130987 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131010 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131036 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131061 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131084 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131107 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131132 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131158 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131181 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131210 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131235 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131265 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131292 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131335 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131364 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131390 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131417 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131442 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131469 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131498 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131522 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131548 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131575 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131688 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131717 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131744 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131775 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131807 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131835 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131864 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131888 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131914 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131939 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.131966 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132031 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132134 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132188 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132255 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132353 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132384 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132411 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132435 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132463 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132497 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132527 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132556 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132584 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132611 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132725 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132744 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132758 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132779 4858 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132794 4858 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132810 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132825 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132840 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132855 4858 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132870 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132884 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132897 4858 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132911 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132926 4858 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132940 4858 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.132956 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.138713 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127893 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127953 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.140652 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126893 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126897 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126919 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126968 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127036 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.140839 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127117 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127464 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.127964 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128000 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128038 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128112 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128299 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.128991 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129041 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.129052 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.134119 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.134524 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135127 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135230 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135261 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135433 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135296 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135459 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135531 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135222 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135853 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.135945 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.136163 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.136889 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.136869 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.137359 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.137402 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.138917 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.139965 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.140244 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.140289 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.140607 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.140933 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.141406 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.141682 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.141727 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.141817 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.141861 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.142023 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.142112 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.142155 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.142410 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.142602 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.142671 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.142985 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.143073 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.143343 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.143401 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.143749 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.143861 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.143979 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.144042 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.126868 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.144395 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.144520 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.145011 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.146596 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.146700 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:23.646661489 +0000 UTC m=+84.967079706 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.147564 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.148078 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.148118 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.148434 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.148781 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.148818 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.148890 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.149091 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.149210 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.145684 4858 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.149756 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.149900 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.150440 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.151057 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.146704 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.145175 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.145432 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.145490 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.146122 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.146249 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.147006 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.145014 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.152475 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.152950 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.155119 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.155670 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.175569 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.161998 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.155532 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.156592 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.159701 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.159721 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.175579 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.160306 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.160340 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.160977 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.161243 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.161267 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.163305 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.163577 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.163622 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.164590 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.175148 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.176000 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.175637 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:23.675607076 +0000 UTC m=+84.996025313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.176021 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.175726 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.176090 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:23.676070688 +0000 UTC m=+84.996488895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.176452 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.176562 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.177271 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.178128 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.178671 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.182476 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.182577 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.182649 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.182758 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:23.682741361 +0000 UTC m=+85.003159558 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.183234 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.184522 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.185022 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.186450 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.188240 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.188461 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.188494 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.188625 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.188895 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.189039 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.189452 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.189684 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.189677 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.190122 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.190220 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.190264 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.190351 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.190441 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.190514 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.190765 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.191022 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.191357 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.191764 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.191984 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.192146 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.192233 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.192781 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.192906 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.193087 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.194144 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.194461 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.194602 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.194624 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.196999 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.198445 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.198063 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.198182 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.198501 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.198584 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.198661 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.198707 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.199517 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.199541 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.199548 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.199595 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.199560 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.201113 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.203402 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.203489 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.204654 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.204876 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.205473 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.205535 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.206003 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.206249 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.206238 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.206686 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.206848 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.206878 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.207141 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.207420 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.209385 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.221744 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.222557 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.233986 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234069 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234158 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234257 4858 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234286 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234281 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234301 4858 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234382 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234400 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234415 4858 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234428 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234442 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234454 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234467 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234479 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234491 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234503 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234520 4858 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234539 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234551 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234564 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234576 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234588 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234600 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234615 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234626 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234638 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234651 4858 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234666 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234678 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234690 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234702 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234713 4858 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234725 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234736 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234748 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234762 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234774 4858 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234787 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234800 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234815 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234827 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234842 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234856 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234868 4858 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234880 4858 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234892 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234904 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234916 4858 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234928 4858 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234940 4858 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234953 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234966 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234978 4858 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.234990 4858 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235002 4858 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235014 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235026 4858 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235039 4858 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235055 4858 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235068 4858 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235081 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235095 4858 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235109 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235121 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235134 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235146 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235157 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235169 4858 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235189 4858 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235201 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235213 4858 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235226 4858 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235238 4858 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235250 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235262 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235273 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235285 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235297 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235308 4858 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235336 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235348 4858 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235359 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235371 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235382 4858 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235393 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235404 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235415 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235426 4858 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235437 4858 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235448 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235459 4858 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235470 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235482 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235495 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235507 4858 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235519 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235531 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235542 4858 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235553 4858 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235567 4858 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235579 4858 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235590 4858 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235608 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235619 4858 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235630 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235642 4858 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235655 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235666 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235680 4858 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235692 4858 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235704 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235715 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235727 4858 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235739 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235750 4858 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235762 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235773 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235784 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235796 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235808 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235824 4858 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235837 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235849 4858 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235860 4858 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235872 4858 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235884 4858 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235895 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235907 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235919 4858 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235931 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235942 4858 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235954 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235966 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235978 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.235989 4858 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236001 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236012 4858 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236025 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236036 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236048 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236059 4858 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236070 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236081 4858 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236092 4858 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236103 4858 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236114 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236126 4858 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236137 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236149 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236160 4858 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236171 4858 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236182 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236194 4858 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236205 4858 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236216 4858 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236228 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236266 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236277 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236288 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236300 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236311 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236339 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236351 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236365 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236377 4858 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236390 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236403 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236414 4858 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236426 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236438 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236449 4858 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236461 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236472 4858 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236483 4858 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236494 4858 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236505 4858 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236516 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236527 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236541 4858 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236552 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.236563 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.367930 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.384650 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.389416 4858 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:23 crc kubenswrapper[4858]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:23 crc kubenswrapper[4858]: set -o allexport Mar 20 08:58:23 crc kubenswrapper[4858]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 20 08:58:23 crc kubenswrapper[4858]: source /etc/kubernetes/apiserver-url.env Mar 20 08:58:23 crc kubenswrapper[4858]: else Mar 20 08:58:23 crc kubenswrapper[4858]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 20 08:58:23 crc kubenswrapper[4858]: exit 1 Mar 20 08:58:23 crc kubenswrapper[4858]: fi Mar 20 08:58:23 crc kubenswrapper[4858]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 20 08:58:23 crc kubenswrapper[4858]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:23 crc kubenswrapper[4858]: > logger="UnhandledError" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.390592 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.391770 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 20 08:58:23 crc kubenswrapper[4858]: W0320 08:58:23.394547 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-a0e04b9e4078a22f06b14b31ee0a6f7aa2570174ee4ed44aeb17a01e15cb27c6 WatchSource:0}: Error finding container a0e04b9e4078a22f06b14b31ee0a6f7aa2570174ee4ed44aeb17a01e15cb27c6: Status 404 returned error can't find the container with id a0e04b9e4078a22f06b14b31ee0a6f7aa2570174ee4ed44aeb17a01e15cb27c6 Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.399412 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.400958 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 20 08:58:23 crc kubenswrapper[4858]: W0320 08:58:23.404948 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-1476fb72be52a88f3cc30ae2ec21600bf0783ee961bd74c36782a370c6424c8c WatchSource:0}: Error finding container 1476fb72be52a88f3cc30ae2ec21600bf0783ee961bd74c36782a370c6424c8c: Status 404 returned error can't find the container with id 1476fb72be52a88f3cc30ae2ec21600bf0783ee961bd74c36782a370c6424c8c Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.407636 4858 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:23 crc kubenswrapper[4858]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:23 crc kubenswrapper[4858]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:23 crc kubenswrapper[4858]: set -o allexport Mar 20 08:58:23 crc kubenswrapper[4858]: source "/env/_master" Mar 20 08:58:23 crc kubenswrapper[4858]: set +o allexport Mar 20 08:58:23 crc kubenswrapper[4858]: fi Mar 20 08:58:23 crc kubenswrapper[4858]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 20 08:58:23 crc kubenswrapper[4858]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 20 08:58:23 crc kubenswrapper[4858]: ho_enable="--enable-hybrid-overlay" Mar 20 08:58:23 crc kubenswrapper[4858]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 20 08:58:23 crc kubenswrapper[4858]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 20 08:58:23 crc kubenswrapper[4858]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 20 08:58:23 crc kubenswrapper[4858]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:23 crc kubenswrapper[4858]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 20 08:58:23 crc kubenswrapper[4858]: --webhook-host=127.0.0.1 \ Mar 20 08:58:23 crc kubenswrapper[4858]: --webhook-port=9743 \ Mar 20 08:58:23 crc kubenswrapper[4858]: ${ho_enable} \ Mar 20 08:58:23 crc kubenswrapper[4858]: --enable-interconnect \ Mar 20 08:58:23 crc kubenswrapper[4858]: --disable-approver \ Mar 20 08:58:23 crc kubenswrapper[4858]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 20 08:58:23 crc kubenswrapper[4858]: --wait-for-kubernetes-api=200s \ Mar 20 08:58:23 crc kubenswrapper[4858]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 20 08:58:23 crc kubenswrapper[4858]: --loglevel="${LOGLEVEL}" Mar 20 08:58:23 crc kubenswrapper[4858]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:23 crc kubenswrapper[4858]: > logger="UnhandledError" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.411348 4858 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:23 crc kubenswrapper[4858]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:23 crc kubenswrapper[4858]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:23 crc kubenswrapper[4858]: set -o allexport Mar 20 08:58:23 crc kubenswrapper[4858]: source "/env/_master" Mar 20 08:58:23 crc kubenswrapper[4858]: set +o allexport Mar 20 08:58:23 crc kubenswrapper[4858]: fi Mar 20 08:58:23 crc kubenswrapper[4858]: Mar 20 08:58:23 crc kubenswrapper[4858]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 20 08:58:23 crc kubenswrapper[4858]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:23 crc kubenswrapper[4858]: --disable-webhook \ Mar 20 08:58:23 crc kubenswrapper[4858]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 20 08:58:23 crc kubenswrapper[4858]: --loglevel="${LOGLEVEL}" Mar 20 08:58:23 crc kubenswrapper[4858]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:23 crc kubenswrapper[4858]: > logger="UnhandledError" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.413143 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.639731 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.639957 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:24.639930575 +0000 UTC m=+85.960348802 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.740190 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.740252 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.740294 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:23 crc kubenswrapper[4858]: I0320 08:58:23.740359 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.740493 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.740596 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:24.740578663 +0000 UTC m=+86.060996880 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.740951 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741007 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:24.740992806 +0000 UTC m=+86.061411013 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741094 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741112 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741125 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741156 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:24.74114596 +0000 UTC m=+86.061564177 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741210 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741222 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741231 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:23 crc kubenswrapper[4858]: E0320 08:58:23.741255 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:24.741247353 +0000 UTC m=+86.061665560 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.073423 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.074471 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.076811 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.078033 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.079971 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.081036 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.082206 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.084106 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.085393 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.086929 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.087449 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.088896 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.089497 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.090024 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.091073 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.091593 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.092604 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.092986 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.093613 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.094847 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.095270 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.096196 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.096619 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.097725 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.098147 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.098779 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.099949 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.100535 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.101694 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.102184 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.103233 4858 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.103376 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.105091 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.106043 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.106503 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.108212 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.108888 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.109762 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.110435 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.111463 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.111907 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.112856 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.113466 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.114599 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.115031 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.115881 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.116446 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.117614 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.118118 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.119136 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.119764 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.120704 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.121297 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.121844 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.210303 4858 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.212154 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.212254 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.212276 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.212426 4858 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.220694 4858 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.221243 4858 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.222777 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.222807 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.222818 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.222835 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.222847 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.239888 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.243201 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.243390 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.243479 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.243540 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.243596 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.255648 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.259034 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.259076 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.259088 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.259103 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.259114 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.267793 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.272081 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.272260 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.272431 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.272555 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.272667 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.282379 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.289839 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.289886 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.289903 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.289936 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.289952 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.304677 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.304823 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.306176 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.306204 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.306215 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.306232 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.306243 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.373941 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c925443184586da9f396891437ed30b08ccdbbfa1b6e10812c05710d9d2c2241"} Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.375075 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1476fb72be52a88f3cc30ae2ec21600bf0783ee961bd74c36782a370c6424c8c"} Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.376425 4858 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:24 crc kubenswrapper[4858]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:24 crc kubenswrapper[4858]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:24 crc kubenswrapper[4858]: set -o allexport Mar 20 08:58:24 crc kubenswrapper[4858]: source "/env/_master" Mar 20 08:58:24 crc kubenswrapper[4858]: set +o allexport Mar 20 08:58:24 crc kubenswrapper[4858]: fi Mar 20 08:58:24 crc kubenswrapper[4858]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 20 08:58:24 crc kubenswrapper[4858]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 20 08:58:24 crc kubenswrapper[4858]: ho_enable="--enable-hybrid-overlay" Mar 20 08:58:24 crc kubenswrapper[4858]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 20 08:58:24 crc kubenswrapper[4858]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 20 08:58:24 crc kubenswrapper[4858]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 20 08:58:24 crc kubenswrapper[4858]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:24 crc kubenswrapper[4858]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 20 08:58:24 crc kubenswrapper[4858]: --webhook-host=127.0.0.1 \ Mar 20 08:58:24 crc kubenswrapper[4858]: --webhook-port=9743 \ Mar 20 08:58:24 crc kubenswrapper[4858]: ${ho_enable} \ Mar 20 08:58:24 crc kubenswrapper[4858]: --enable-interconnect \ Mar 20 08:58:24 crc kubenswrapper[4858]: --disable-approver \ Mar 20 08:58:24 crc kubenswrapper[4858]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 20 08:58:24 crc kubenswrapper[4858]: --wait-for-kubernetes-api=200s \ Mar 20 08:58:24 crc kubenswrapper[4858]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 20 08:58:24 crc kubenswrapper[4858]: --loglevel="${LOGLEVEL}" Mar 20 08:58:24 crc kubenswrapper[4858]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:24 crc kubenswrapper[4858]: > logger="UnhandledError" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.376526 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"a0e04b9e4078a22f06b14b31ee0a6f7aa2570174ee4ed44aeb17a01e15cb27c6"} Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.377628 4858 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:24 crc kubenswrapper[4858]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 20 08:58:24 crc kubenswrapper[4858]: set -o allexport Mar 20 08:58:24 crc kubenswrapper[4858]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 20 08:58:24 crc kubenswrapper[4858]: source /etc/kubernetes/apiserver-url.env Mar 20 08:58:24 crc kubenswrapper[4858]: else Mar 20 08:58:24 crc kubenswrapper[4858]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 20 08:58:24 crc kubenswrapper[4858]: exit 1 Mar 20 08:58:24 crc kubenswrapper[4858]: fi Mar 20 08:58:24 crc kubenswrapper[4858]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 20 08:58:24 crc kubenswrapper[4858]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:24 crc kubenswrapper[4858]: > logger="UnhandledError" Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.378094 4858 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 08:58:24 crc kubenswrapper[4858]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 20 08:58:24 crc kubenswrapper[4858]: if [[ -f "/env/_master" ]]; then Mar 20 08:58:24 crc kubenswrapper[4858]: set -o allexport Mar 20 08:58:24 crc kubenswrapper[4858]: source "/env/_master" Mar 20 08:58:24 crc kubenswrapper[4858]: set +o allexport Mar 20 08:58:24 crc kubenswrapper[4858]: fi Mar 20 08:58:24 crc kubenswrapper[4858]: Mar 20 08:58:24 crc kubenswrapper[4858]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 20 08:58:24 crc kubenswrapper[4858]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 20 08:58:24 crc kubenswrapper[4858]: --disable-webhook \ Mar 20 08:58:24 crc kubenswrapper[4858]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 20 08:58:24 crc kubenswrapper[4858]: --loglevel="${LOGLEVEL}" Mar 20 08:58:24 crc kubenswrapper[4858]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 20 08:58:24 crc kubenswrapper[4858]: > logger="UnhandledError" Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.378483 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.378722 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.379187 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.379913 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.388814 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.403351 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.411046 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.411082 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.411093 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.411107 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.411118 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.418754 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.432203 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.441946 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.451550 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.461658 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.469747 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.482702 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.493689 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.507592 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.513885 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.513943 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.513955 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.513981 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.514000 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.518047 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.616639 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.616720 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.616732 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.616750 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.616760 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.649562 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.649807 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:26.649759589 +0000 UTC m=+87.970177826 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.719821 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.719897 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.719918 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.719945 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.719968 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.751024 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.751109 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.751203 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.751290 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751487 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751512 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751541 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751548 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751642 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:26.7516126 +0000 UTC m=+88.072030837 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751710 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:26.751673332 +0000 UTC m=+88.072091739 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751524 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751768 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751788 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751830 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:26.751820476 +0000 UTC m=+88.072238693 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751582 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:24 crc kubenswrapper[4858]: E0320 08:58:24.751883 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:26.751876357 +0000 UTC m=+88.072294564 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.823925 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.824010 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.824050 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.824081 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.824105 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.926871 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.926968 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.926993 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.927026 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:24 crc kubenswrapper[4858]: I0320 08:58:24.927050 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:24Z","lastTransitionTime":"2026-03-20T08:58:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.029245 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.029303 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.029392 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.029418 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.029438 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.069218 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.069288 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:25 crc kubenswrapper[4858]: E0320 08:58:25.069420 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:25 crc kubenswrapper[4858]: E0320 08:58:25.069562 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.069587 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:25 crc kubenswrapper[4858]: E0320 08:58:25.069727 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.131682 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.131752 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.131771 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.131796 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.131814 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.234769 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.234842 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.234865 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.234895 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.234916 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.337744 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.337809 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.337830 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.337859 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.337882 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.440396 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.440483 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.440511 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.440543 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.440568 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.544350 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.544407 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.544424 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.544446 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.544473 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.646829 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.646857 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.646865 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.646878 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.646886 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.750086 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.750142 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.750160 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.750185 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.750201 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.852807 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.852839 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.852847 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.852859 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.852868 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.956062 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.956120 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.956137 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.956161 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:25 crc kubenswrapper[4858]: I0320 08:58:25.956177 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:25Z","lastTransitionTime":"2026-03-20T08:58:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.065455 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.065523 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.065546 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.065575 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.065596 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.168907 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.168947 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.168958 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.168973 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.168985 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.271639 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.271718 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.271737 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.271765 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.271784 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.375187 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.375274 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.375297 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.375362 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.375386 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.478029 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.478095 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.478113 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.478140 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.478157 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.581108 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.581164 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.581184 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.581212 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.581232 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.669556 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.669823 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:30.669799426 +0000 UTC m=+91.990217663 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.683532 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.683609 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.683633 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.683661 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.683683 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.770081 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.770143 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.770176 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.770206 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770279 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770362 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770386 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770399 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770468 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:30.770396343 +0000 UTC m=+92.090814600 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770475 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770508 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:30.770491115 +0000 UTC m=+92.090909482 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770512 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770607 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:30.770586338 +0000 UTC m=+92.091004525 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770520 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770645 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:26 crc kubenswrapper[4858]: E0320 08:58:26.770673 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:30.77066625 +0000 UTC m=+92.091084447 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.786894 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.786969 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.786993 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.787024 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.787049 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.889407 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.889467 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.889486 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.889509 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.889527 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.992162 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.992211 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.992233 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.992251 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:26 crc kubenswrapper[4858]: I0320 08:58:26.992263 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:26Z","lastTransitionTime":"2026-03-20T08:58:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.069552 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.069572 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:27 crc kubenswrapper[4858]: E0320 08:58:27.069668 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.069690 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:27 crc kubenswrapper[4858]: E0320 08:58:27.069868 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:27 crc kubenswrapper[4858]: E0320 08:58:27.069979 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.093899 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.093960 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.093981 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.094006 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.094025 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.196578 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.196610 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.196617 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.196631 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.196640 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.299736 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.299795 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.299809 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.299829 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.299844 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.402442 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.402500 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.402513 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.402528 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.402537 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.505412 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.505466 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.505482 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.505503 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.505518 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.607971 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.608047 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.608061 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.608118 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.608134 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.711094 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.711136 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.711145 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.711157 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.711166 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.812856 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.812911 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.812923 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.812941 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.812953 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.915659 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.915696 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.915705 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.915720 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:27 crc kubenswrapper[4858]: I0320 08:58:27.915730 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:27Z","lastTransitionTime":"2026-03-20T08:58:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.017698 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.017743 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.017754 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.017771 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.017782 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.120243 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.120303 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.120352 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.120376 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.120391 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.223449 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.223566 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.223589 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.223629 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.223653 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.327556 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.327651 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.327673 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.327705 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.327727 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.430850 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.430932 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.430945 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.430969 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.430979 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.533662 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.533712 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.533721 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.533736 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.533747 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.636099 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.636158 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.636176 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.636200 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.636221 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.739419 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.739485 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.739498 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.739524 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.739540 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.843503 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.843589 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.843608 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.843635 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.843658 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.947135 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.947182 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.947195 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.947220 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:28 crc kubenswrapper[4858]: I0320 08:58:28.947231 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:28Z","lastTransitionTime":"2026-03-20T08:58:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.050166 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.050226 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.050237 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.050261 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.050275 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.069913 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.069995 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.070055 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:29 crc kubenswrapper[4858]: E0320 08:58:29.070098 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:29 crc kubenswrapper[4858]: E0320 08:58:29.070180 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:29 crc kubenswrapper[4858]: E0320 08:58:29.070238 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.153443 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.153501 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.153514 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.153536 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.153547 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.256144 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.256212 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.256227 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.256249 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.256263 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.301424 4858 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.359755 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.359825 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.359837 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.359862 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.359880 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.463059 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.463135 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.463151 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.463174 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.463190 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.566933 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.567035 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.567061 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.567404 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.567441 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.671174 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.671243 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.671263 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.671297 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.671362 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.774094 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.774147 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.774163 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.774188 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.774205 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.877350 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.877728 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.877849 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.877949 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.878014 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.980455 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.980496 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.980508 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.980555 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:29 crc kubenswrapper[4858]: I0320 08:58:29.980566 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:29Z","lastTransitionTime":"2026-03-20T08:58:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.080880 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.082019 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.082190 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.082415 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.082571 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.082684 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.091816 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.102069 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.109661 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.117487 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.124633 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.184180 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.184213 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.184221 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.184235 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.184244 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.286980 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.287595 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.287681 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.287761 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.287847 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.390080 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.390123 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.390138 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.390158 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.390173 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.492217 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.492273 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.492290 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.492358 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.492396 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.595104 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.595180 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.595203 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.595231 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.595254 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.698050 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.698104 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.698121 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.698141 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.698156 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.707417 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.707565 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:38.707548446 +0000 UTC m=+100.027966643 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.800877 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.800933 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.800944 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.800961 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.800973 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.808892 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.808947 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.808974 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.809002 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809117 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809117 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809127 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809181 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809199 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:38.809177722 +0000 UTC m=+100.129595939 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809212 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809227 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:38.809214393 +0000 UTC m=+100.129632610 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809230 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809137 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809253 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809281 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:38.809264975 +0000 UTC m=+100.129683182 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:30 crc kubenswrapper[4858]: E0320 08:58:30.809337 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:38.809290675 +0000 UTC m=+100.129708892 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.903593 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.903627 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.903637 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.903651 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:30 crc kubenswrapper[4858]: I0320 08:58:30.903662 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:30Z","lastTransitionTime":"2026-03-20T08:58:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.006385 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.006427 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.006447 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.006462 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.006471 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.070105 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.070148 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:31 crc kubenswrapper[4858]: E0320 08:58:31.070200 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.070363 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:31 crc kubenswrapper[4858]: E0320 08:58:31.070663 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:31 crc kubenswrapper[4858]: E0320 08:58:31.070846 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.083221 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.083268 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.083503 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 08:58:31 crc kubenswrapper[4858]: E0320 08:58:31.083674 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.108253 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.108456 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.108551 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.108629 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.108689 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.210994 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.211022 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.211031 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.211044 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.211052 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.314917 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.314946 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.314956 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.314970 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.314980 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.394391 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 08:58:31 crc kubenswrapper[4858]: E0320 08:58:31.394567 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.417347 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.417398 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.417413 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.417431 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.417445 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.520110 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.520177 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.520200 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.520223 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.520240 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.622378 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.622447 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.622466 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.622487 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.622503 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.724494 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.724539 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.724551 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.724568 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.724578 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.827292 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.827411 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.827439 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.827466 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.827483 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.929943 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.930014 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.930052 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.930087 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:31 crc kubenswrapper[4858]: I0320 08:58:31.930109 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:31Z","lastTransitionTime":"2026-03-20T08:58:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.032947 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.032985 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.032996 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.033019 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.033034 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.134672 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.135030 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.135041 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.135059 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.135071 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.237992 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.238047 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.238059 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.238078 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.238091 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.340662 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.340726 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.340769 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.340810 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.340838 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.443494 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.443549 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.443558 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.443572 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.443581 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.546330 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.546367 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.546378 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.546394 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.546404 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.648841 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.649121 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.649300 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.649441 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.649541 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.751574 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.751620 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.751630 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.751645 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.751655 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.854822 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.854879 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.854908 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.854936 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.854955 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.957711 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.958023 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.958122 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.958251 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:32 crc kubenswrapper[4858]: I0320 08:58:32.958378 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:32Z","lastTransitionTime":"2026-03-20T08:58:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.060907 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.060946 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.060955 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.060969 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.060979 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.069557 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:33 crc kubenswrapper[4858]: E0320 08:58:33.069677 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.069986 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:33 crc kubenswrapper[4858]: E0320 08:58:33.070052 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.070093 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:33 crc kubenswrapper[4858]: E0320 08:58:33.070141 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.163826 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.164199 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.164430 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.164698 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.164885 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.267859 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.267905 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.267913 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.267937 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.267950 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.370370 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.370982 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.371134 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.371258 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.371394 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.474784 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.474849 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.474873 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.474905 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.474931 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.577947 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.578116 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.578147 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.578180 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.578205 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.680096 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.680355 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.680500 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.680568 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.680763 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.784099 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.784410 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.784500 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.784595 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.784688 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.887358 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.887395 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.887410 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.887429 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.887442 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.990198 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.990247 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.990256 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.990271 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:33 crc kubenswrapper[4858]: I0320 08:58:33.990280 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:33Z","lastTransitionTime":"2026-03-20T08:58:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.092295 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.092553 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.092617 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.092708 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.092764 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.194702 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.195004 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.195081 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.195161 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.195268 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.297891 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.297951 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.297972 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.298003 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.298024 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.405419 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.405456 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.405470 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.405488 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.405499 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.508859 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.508920 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.508936 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.508957 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.508973 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.590912 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.590999 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.591020 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.591052 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.591077 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: E0320 08:58:34.605111 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.611648 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.611796 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.612178 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.612469 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.612740 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.617264 4858 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 08:58:34 crc kubenswrapper[4858]: E0320 08:58:34.627252 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.631251 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.631291 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.631303 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.631346 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.631358 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: E0320 08:58:34.641730 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.645192 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.645234 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.645246 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.645261 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.645273 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: E0320 08:58:34.655699 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.660452 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.660526 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.660538 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.660558 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.660594 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: E0320 08:58:34.670086 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:34 crc kubenswrapper[4858]: E0320 08:58:34.670251 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.672832 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.672903 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.672931 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.672964 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.672993 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.775490 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.775534 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.775545 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.775561 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.775573 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.878950 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.878997 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.879009 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.879025 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.879038 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.981910 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.981967 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.981981 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.982000 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:34 crc kubenswrapper[4858]: I0320 08:58:34.982018 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:34Z","lastTransitionTime":"2026-03-20T08:58:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.069001 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.069001 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:35 crc kubenswrapper[4858]: E0320 08:58:35.069133 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:35 crc kubenswrapper[4858]: E0320 08:58:35.069200 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.069001 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:35 crc kubenswrapper[4858]: E0320 08:58:35.069303 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.085391 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.085434 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.085442 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.085456 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.085466 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.187445 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.187538 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.187547 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.187565 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.187575 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.290873 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.290919 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.290929 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.290943 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.290952 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.394014 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.394079 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.394089 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.394111 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.394123 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.497747 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.497821 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.497845 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.497876 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.497899 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.601279 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.601451 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.601476 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.601545 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.601571 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.703897 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.703950 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.703973 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.703999 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.704014 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.808152 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.808218 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.808246 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.808277 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.808301 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.911505 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.911574 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.911594 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.911621 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:35 crc kubenswrapper[4858]: I0320 08:58:35.911639 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:35Z","lastTransitionTime":"2026-03-20T08:58:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.013416 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.013482 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.013494 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.013506 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.013514 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.115999 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.116086 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.116097 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.116113 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.116126 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.219346 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.219641 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.219655 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.219671 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.219684 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.322134 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.322195 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.322207 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.322226 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.322236 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.425307 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.425410 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.425429 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.425468 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.425490 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.528125 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.528180 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.528192 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.528211 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.528224 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.630921 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.630962 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.630972 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.630989 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.631001 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.734229 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.734278 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.734289 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.734304 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.734357 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.836645 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.836703 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.836715 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.836734 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.836747 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.939845 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.940115 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.940251 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.940331 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.940389 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:36Z","lastTransitionTime":"2026-03-20T08:58:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.982710 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-mwh2v"] Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.983251 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.985079 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.985687 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.986998 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 08:58:36 crc kubenswrapper[4858]: I0320 08:58:36.997586 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.010125 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.022755 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.032900 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.039864 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.042549 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.042588 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.042597 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.042612 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.042622 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:37Z","lastTransitionTime":"2026-03-20T08:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.051845 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.064120 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.068064 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q64d\" (UniqueName: \"kubernetes.io/projected/da8fcd26-7c6c-4a53-a6c6-dadde0238068-kube-api-access-8q64d\") pod \"node-resolver-mwh2v\" (UID: \"da8fcd26-7c6c-4a53-a6c6-dadde0238068\") " pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.068107 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/da8fcd26-7c6c-4a53-a6c6-dadde0238068-hosts-file\") pod \"node-resolver-mwh2v\" (UID: \"da8fcd26-7c6c-4a53-a6c6-dadde0238068\") " pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.069344 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:37 crc kubenswrapper[4858]: E0320 08:58:37.069427 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.069475 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:37 crc kubenswrapper[4858]: E0320 08:58:37.069567 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.069791 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:37 crc kubenswrapper[4858]: E0320 08:58:37.069986 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.075789 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.082185 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.146093 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.146161 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.146173 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.146192 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.146204 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:37Z","lastTransitionTime":"2026-03-20T08:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.168737 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q64d\" (UniqueName: \"kubernetes.io/projected/da8fcd26-7c6c-4a53-a6c6-dadde0238068-kube-api-access-8q64d\") pod \"node-resolver-mwh2v\" (UID: \"da8fcd26-7c6c-4a53-a6c6-dadde0238068\") " pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.168817 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/da8fcd26-7c6c-4a53-a6c6-dadde0238068-hosts-file\") pod \"node-resolver-mwh2v\" (UID: \"da8fcd26-7c6c-4a53-a6c6-dadde0238068\") " pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.168923 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/da8fcd26-7c6c-4a53-a6c6-dadde0238068-hosts-file\") pod \"node-resolver-mwh2v\" (UID: \"da8fcd26-7c6c-4a53-a6c6-dadde0238068\") " pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.202448 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q64d\" (UniqueName: \"kubernetes.io/projected/da8fcd26-7c6c-4a53-a6c6-dadde0238068-kube-api-access-8q64d\") pod \"node-resolver-mwh2v\" (UID: \"da8fcd26-7c6c-4a53-a6c6-dadde0238068\") " pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.249760 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.249837 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.249871 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.249904 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.249926 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:37Z","lastTransitionTime":"2026-03-20T08:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.306822 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mwh2v" Mar 20 08:58:37 crc kubenswrapper[4858]: W0320 08:58:37.317120 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda8fcd26_7c6c_4a53_a6c6_dadde0238068.slice/crio-27b446f6ceb3a4244f5fbefba52328e2c2e096be0500198799f0311d412243d3 WatchSource:0}: Error finding container 27b446f6ceb3a4244f5fbefba52328e2c2e096be0500198799f0311d412243d3: Status 404 returned error can't find the container with id 27b446f6ceb3a4244f5fbefba52328e2c2e096be0500198799f0311d412243d3 Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.352422 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.352457 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.352466 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.352479 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.352488 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:37Z","lastTransitionTime":"2026-03-20T08:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.370470 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-t45zv"] Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.371290 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-w6t79"] Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.371534 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-p2cjs"] Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.371751 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.371768 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.371768 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.373576 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.374229 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.374487 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.374881 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.374940 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.375512 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.375993 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.376041 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.376095 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.376042 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.376287 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.376481 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.381207 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.391786 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.401008 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.409907 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.412834 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mwh2v" event={"ID":"da8fcd26-7c6c-4a53-a6c6-dadde0238068","Type":"ContainerStarted","Data":"27b446f6ceb3a4244f5fbefba52328e2c2e096be0500198799f0311d412243d3"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.424205 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.433903 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.444104 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.452468 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.454049 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.454070 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.454078 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.454091 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.454099 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:37Z","lastTransitionTime":"2026-03-20T08:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.464971 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471586 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24656c62-314b-4c20-adf1-217d58a95f57-cni-binary-copy\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471622 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/584bd2e0-0786-4137-9674-790c8fb680c5-proxy-tls\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471683 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-netns\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471737 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znr6q\" (UniqueName: \"kubernetes.io/projected/584bd2e0-0786-4137-9674-790c8fb680c5-kube-api-access-znr6q\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471765 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-etc-kubernetes\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471781 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-cnibin\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471798 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-multus-certs\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471837 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/584bd2e0-0786-4137-9674-790c8fb680c5-rootfs\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471858 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-cni-bin\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471886 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-hostroot\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471920 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-os-release\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471953 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-cnibin\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.471972 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-system-cni-dir\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472006 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/24656c62-314b-4c20-adf1-217d58a95f57-multus-daemon-config\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472040 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpfrn\" (UniqueName: \"kubernetes.io/projected/e28e3e9c-e621-4e85-af97-c3f48adb269d-kube-api-access-fpfrn\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472066 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-os-release\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472080 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e28e3e9c-e621-4e85-af97-c3f48adb269d-cni-binary-copy\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472096 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-cni-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472112 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-socket-dir-parent\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472154 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/584bd2e0-0786-4137-9674-790c8fb680c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472173 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-cni-multus\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472186 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-conf-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472224 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472251 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-kubelet\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472265 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e28e3e9c-e621-4e85-af97-c3f48adb269d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472278 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-system-cni-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472293 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcxb2\" (UniqueName: \"kubernetes.io/projected/24656c62-314b-4c20-adf1-217d58a95f57-kube-api-access-hcxb2\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472277 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.472309 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-k8s-cni-cncf-io\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.480789 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.486580 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.496102 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.503940 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.511865 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.520797 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.532549 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.543333 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.555027 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.556004 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.556024 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.556033 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.556048 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.556059 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:37Z","lastTransitionTime":"2026-03-20T08:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.565090 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573342 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573390 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-kubelet\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573409 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcxb2\" (UniqueName: \"kubernetes.io/projected/24656c62-314b-4c20-adf1-217d58a95f57-kube-api-access-hcxb2\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573426 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e28e3e9c-e621-4e85-af97-c3f48adb269d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573446 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-system-cni-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573462 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-k8s-cni-cncf-io\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573489 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24656c62-314b-4c20-adf1-217d58a95f57-cni-binary-copy\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573507 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/584bd2e0-0786-4137-9674-790c8fb680c5-proxy-tls\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573524 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-netns\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573541 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znr6q\" (UniqueName: \"kubernetes.io/projected/584bd2e0-0786-4137-9674-790c8fb680c5-kube-api-access-znr6q\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573557 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-etc-kubernetes\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573573 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-cnibin\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573589 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-multus-certs\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573612 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/584bd2e0-0786-4137-9674-790c8fb680c5-rootfs\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573627 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-cni-bin\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573642 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-os-release\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573659 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-hostroot\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573679 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-cnibin\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573699 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-system-cni-dir\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573714 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-os-release\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573729 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/24656c62-314b-4c20-adf1-217d58a95f57-multus-daemon-config\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573750 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpfrn\" (UniqueName: \"kubernetes.io/projected/e28e3e9c-e621-4e85-af97-c3f48adb269d-kube-api-access-fpfrn\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573765 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e28e3e9c-e621-4e85-af97-c3f48adb269d-cni-binary-copy\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573782 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-cni-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573798 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-socket-dir-parent\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573846 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/584bd2e0-0786-4137-9674-790c8fb680c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573861 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-cni-multus\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573877 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-conf-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573919 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/584bd2e0-0786-4137-9674-790c8fb680c5-rootfs\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573957 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-cni-bin\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573936 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-conf-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.573980 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-kubelet\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574017 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-os-release\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574044 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-hostroot\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574075 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-cnibin\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574100 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-k8s-cni-cncf-io\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574169 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-system-cni-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574197 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-system-cni-dir\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574231 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-os-release\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574248 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574296 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-etc-kubernetes\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574607 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-socket-dir-parent\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.575146 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e28e3e9c-e621-4e85-af97-c3f48adb269d-cnibin\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.575218 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-multus-cni-dir\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.575257 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-var-lib-cni-multus\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.575300 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24656c62-314b-4c20-adf1-217d58a95f57-cni-binary-copy\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.574073 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-netns\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.575373 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/24656c62-314b-4c20-adf1-217d58a95f57-host-run-multus-certs\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.575649 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/24656c62-314b-4c20-adf1-217d58a95f57-multus-daemon-config\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.575821 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/584bd2e0-0786-4137-9674-790c8fb680c5-mcd-auth-proxy-config\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.576201 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e28e3e9c-e621-4e85-af97-c3f48adb269d-cni-binary-copy\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.577172 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e28e3e9c-e621-4e85-af97-c3f48adb269d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.578244 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/584bd2e0-0786-4137-9674-790c8fb680c5-proxy-tls\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.585734 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.589294 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znr6q\" (UniqueName: \"kubernetes.io/projected/584bd2e0-0786-4137-9674-790c8fb680c5-kube-api-access-znr6q\") pod \"machine-config-daemon-w6t79\" (UID: \"584bd2e0-0786-4137-9674-790c8fb680c5\") " pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.590780 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcxb2\" (UniqueName: \"kubernetes.io/projected/24656c62-314b-4c20-adf1-217d58a95f57-kube-api-access-hcxb2\") pod \"multus-p2cjs\" (UID: \"24656c62-314b-4c20-adf1-217d58a95f57\") " pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.593994 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpfrn\" (UniqueName: \"kubernetes.io/projected/e28e3e9c-e621-4e85-af97-c3f48adb269d-kube-api-access-fpfrn\") pod \"multus-additional-cni-plugins-t45zv\" (UID: \"e28e3e9c-e621-4e85-af97-c3f48adb269d\") " pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.600427 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.611900 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.889126 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-p2cjs" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.889256 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.889506 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-t45zv" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.898219 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.898263 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.898276 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.898294 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.898306 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:37Z","lastTransitionTime":"2026-03-20T08:58:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.910121 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dwpzf"] Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.911430 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.916379 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.916927 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.917114 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.918496 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.918616 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.918772 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.919094 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.931079 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.940124 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.949513 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.961510 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.971655 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.979648 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.987991 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:37 crc kubenswrapper[4858]: I0320 08:58:37.996020 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.000390 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.000431 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.000442 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.000460 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.000470 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.005449 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.015963 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.025428 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.036353 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.055538 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.091116 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-netns\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.091163 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-slash\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.091191 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-ovn-kubernetes\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095769 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-node-log\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095804 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-netd\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095827 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-ovn\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095856 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-bin\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095880 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095900 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095936 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-kubelet\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095961 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-systemd-units\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.095992 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-var-lib-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096028 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-script-lib\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096106 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovn-node-metrics-cert\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096181 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-log-socket\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096208 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-config\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096233 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-env-overrides\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096295 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htvhn\" (UniqueName: \"kubernetes.io/projected/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-kube-api-access-htvhn\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096348 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-systemd\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.096380 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-etc-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.103639 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.103690 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.103707 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.103738 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.103756 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196775 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-kubelet\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196806 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-systemd-units\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196821 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-var-lib-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196838 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-script-lib\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196862 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-log-socket\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196878 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-config\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196892 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-env-overrides\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196892 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-kubelet\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196905 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovn-node-metrics-cert\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.196996 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-systemd\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197021 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htvhn\" (UniqueName: \"kubernetes.io/projected/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-kube-api-access-htvhn\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197066 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-etc-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197092 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-slash\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197120 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-netns\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197161 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-ovn-kubernetes\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197268 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-var-lib-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197326 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-systemd-units\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197452 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-log-socket\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197585 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-netns\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197613 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-etc-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.197646 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-slash\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198049 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-config\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198075 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-systemd\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198083 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-script-lib\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198124 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-env-overrides\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198406 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-ovn-kubernetes\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198667 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-node-log\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198734 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-netd\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198771 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-ovn\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198787 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-node-log\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198801 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-bin\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198818 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-netd\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198837 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198846 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-ovn\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198869 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198918 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198935 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-bin\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.198970 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-openvswitch\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.201257 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovn-node-metrics-cert\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.206715 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.206745 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.206754 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.206771 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.206780 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.225864 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htvhn\" (UniqueName: \"kubernetes.io/projected/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-kube-api-access-htvhn\") pod \"ovnkube-node-dwpzf\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.234264 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:38 crc kubenswrapper[4858]: W0320 08:58:38.247272 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21fd7c33_ddc7_4a05_a922_472eb8ccd4e1.slice/crio-cfff7b09664ee0747c5d415bfaf39d57445ed67a89eea47f386ef37965a3205e WatchSource:0}: Error finding container cfff7b09664ee0747c5d415bfaf39d57445ed67a89eea47f386ef37965a3205e: Status 404 returned error can't find the container with id cfff7b09664ee0747c5d415bfaf39d57445ed67a89eea47f386ef37965a3205e Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.309529 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.309571 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.309583 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.309602 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.309615 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.412262 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.412292 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.412302 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.412327 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.412337 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.417054 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.420101 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mwh2v" event={"ID":"da8fcd26-7c6c-4a53-a6c6-dadde0238068","Type":"ContainerStarted","Data":"1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.421896 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a" exitCode=0 Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.421964 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.421984 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"cfff7b09664ee0747c5d415bfaf39d57445ed67a89eea47f386ef37965a3205e"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.423283 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerStarted","Data":"574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.423334 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerStarted","Data":"c3ebe7b483626a8fef4a7109d8c8b225a30bcc8459bf2cb2c49dd2748e77c135"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.425731 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.425777 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.425790 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"a0caeadd33786bc85f29016374a205e9b820e8d8cf9dd3fce5e22e56c30609d3"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.427476 4858 generic.go:334] "Generic (PLEG): container finished" podID="e28e3e9c-e621-4e85-af97-c3f48adb269d" containerID="cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d" exitCode=0 Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.427600 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerDied","Data":"cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.427703 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerStarted","Data":"930d4668a20b64a06e83ff152f0b2cbc933ed66462603c30f84fbe75e1c321af"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.431013 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.433902 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.433952 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.445698 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.452280 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.460456 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.469685 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.477199 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.487200 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.500092 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.512392 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.514412 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.514441 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.514453 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.514467 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.514480 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.523476 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.534657 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.541961 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.549420 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.557199 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.570778 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.586117 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.609285 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.624908 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.624941 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.624952 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.624967 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.624976 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.636944 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.656645 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.670348 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.682679 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.694871 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.717735 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.727067 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.727457 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.727469 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.727487 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.727499 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.731140 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.742556 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.759229 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.803445 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.803666 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.803646646 +0000 UTC m=+116.124064843 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.830686 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.830739 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.830750 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.830771 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.830786 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.904670 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.904735 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.904780 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904835 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904859 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904871 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904913 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.904899031 +0000 UTC m=+116.225317228 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.904836 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904927 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904971 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.905009 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.905024 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904982 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.904965893 +0000 UTC m=+116.225384100 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.904971 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.905111 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.905090496 +0000 UTC m=+116.225508693 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:38 crc kubenswrapper[4858]: E0320 08:58:38.905166 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:58:54.905144608 +0000 UTC m=+116.225562885 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.933091 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.933122 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.933134 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.933151 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:38 crc kubenswrapper[4858]: I0320 08:58:38.933161 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:38Z","lastTransitionTime":"2026-03-20T08:58:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.035034 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.035067 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.035075 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.035089 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.035098 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.069597 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.069723 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.069787 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:39 crc kubenswrapper[4858]: E0320 08:58:39.069918 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:39 crc kubenswrapper[4858]: E0320 08:58:39.070163 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:39 crc kubenswrapper[4858]: E0320 08:58:39.070333 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.137018 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.137056 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.137068 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.137083 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.137094 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.240447 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.240485 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.240494 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.240507 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.240516 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.343089 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.343130 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.343140 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.343155 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.343172 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.440204 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.440250 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.440264 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.440275 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.440287 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.440299 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.441499 4858 generic.go:334] "Generic (PLEG): container finished" podID="e28e3e9c-e621-4e85-af97-c3f48adb269d" containerID="a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002" exitCode=0 Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.441553 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerDied","Data":"a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.445397 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.445434 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.445447 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.445461 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.445473 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.455460 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.469486 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.487721 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.502118 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.516789 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.531567 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.548533 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.553955 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.554005 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.554017 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.554071 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.554088 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.563600 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.577951 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.594265 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.609267 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.620875 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.637213 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.657080 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.657119 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.657131 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.657152 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.657164 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.705787 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-2j2m8"] Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.706394 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.708722 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.708806 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.709055 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.710611 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.710727 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-serviceca\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.710759 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-host\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.711046 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qgs4\" (UniqueName: \"kubernetes.io/projected/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-kube-api-access-4qgs4\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.722688 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.736312 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.759371 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.759634 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.759692 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.759702 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.759725 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.759738 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.777569 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.795185 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.810134 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.811959 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qgs4\" (UniqueName: \"kubernetes.io/projected/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-kube-api-access-4qgs4\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.812021 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-serviceca\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.812053 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-host\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.812135 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-host\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.815198 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-serviceca\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.836236 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.847644 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qgs4\" (UniqueName: \"kubernetes.io/projected/49ca0c4b-5811-4080-b1a6-1c6f02fc5d76-kube-api-access-4qgs4\") pod \"node-ca-2j2m8\" (UID: \"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\") " pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.850726 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.861366 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.861397 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.861405 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.861418 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.861427 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.868779 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.884199 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.905268 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.920181 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.935708 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.950201 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:39Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.964406 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.964444 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.964467 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.964483 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:39 crc kubenswrapper[4858]: I0320 08:58:39.964493 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:39Z","lastTransitionTime":"2026-03-20T08:58:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.056865 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-2j2m8" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.067167 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.067232 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.067251 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.067274 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.067292 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: W0320 08:58:40.074229 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49ca0c4b_5811_4080_b1a6_1c6f02fc5d76.slice/crio-b5d0ef51cc447e23bcbb708d6f8e029ad09318cf420f1bbfd7ae64bcc346f2f0 WatchSource:0}: Error finding container b5d0ef51cc447e23bcbb708d6f8e029ad09318cf420f1bbfd7ae64bcc346f2f0: Status 404 returned error can't find the container with id b5d0ef51cc447e23bcbb708d6f8e029ad09318cf420f1bbfd7ae64bcc346f2f0 Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.087047 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.106646 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.117003 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.127812 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.140836 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.155226 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.172363 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.172403 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.172414 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.172436 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.172450 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.174690 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.189598 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.206751 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.222975 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.242033 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.254980 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.266996 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.275103 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.275145 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.275154 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.275172 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.275183 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.281728 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.378704 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.379253 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.379277 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.379304 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.379363 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.447271 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2j2m8" event={"ID":"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76","Type":"ContainerStarted","Data":"10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.447350 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-2j2m8" event={"ID":"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76","Type":"ContainerStarted","Data":"b5d0ef51cc447e23bcbb708d6f8e029ad09318cf420f1bbfd7ae64bcc346f2f0"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.452276 4858 generic.go:334] "Generic (PLEG): container finished" podID="e28e3e9c-e621-4e85-af97-c3f48adb269d" containerID="388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d" exitCode=0 Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.452396 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerDied","Data":"388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.465933 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.484820 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.487481 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.487529 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.487544 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.487561 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.487573 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.499162 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.513268 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.525835 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.539228 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.569753 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.596891 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.596948 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.596970 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.596987 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.596997 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.604936 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.632159 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.647970 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.664339 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.676623 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.694500 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.699931 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.699980 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.699994 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.700017 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.700034 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.708492 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.726893 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.756166 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.795987 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.802061 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.802114 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.802131 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.802150 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.802164 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.851478 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.874465 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.905185 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.905257 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.905267 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.905294 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.905343 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:40Z","lastTransitionTime":"2026-03-20T08:58:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.916240 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.955917 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:40 crc kubenswrapper[4858]: I0320 08:58:40.995187 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.007968 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.008018 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.008030 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.008049 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.008063 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.035817 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.069463 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.069536 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.069595 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:41 crc kubenswrapper[4858]: E0320 08:58:41.069675 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:41 crc kubenswrapper[4858]: E0320 08:58:41.069784 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:41 crc kubenswrapper[4858]: E0320 08:58:41.069910 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.077790 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.110759 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.110805 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.110838 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.110862 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.110877 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.116644 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.155251 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.194678 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.213443 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.213487 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.213500 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.213519 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.213640 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.235360 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.316721 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.316773 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.316787 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.316813 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.316834 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.419806 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.419871 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.419882 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.419900 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.419912 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.458075 4858 generic.go:334] "Generic (PLEG): container finished" podID="e28e3e9c-e621-4e85-af97-c3f48adb269d" containerID="73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba" exitCode=0 Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.458172 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerDied","Data":"73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.459589 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.465877 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.480504 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.497683 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.512491 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.524934 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.525788 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.525824 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.525835 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.525853 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.525865 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.537830 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.557173 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.575736 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.590228 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.604551 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.628094 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.628130 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.628138 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.628151 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.628160 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.646008 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.674253 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.713404 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.730662 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.730703 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.730715 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.730732 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.730750 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.759715 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.798512 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.835127 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.835192 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.835204 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.835223 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.835236 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.841012 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.876360 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.917015 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.945136 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.945236 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.945252 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.945371 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.945418 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:41Z","lastTransitionTime":"2026-03-20T08:58:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.953865 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:41 crc kubenswrapper[4858]: I0320 08:58:41.999065 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:41Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.031902 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.048551 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.048578 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.048586 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.048598 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.048606 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.073735 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.120551 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.150724 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.150771 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.150783 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.150800 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.150811 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.159004 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.194717 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.236020 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.252977 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.253029 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.253039 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.253055 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.253065 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.281500 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.315143 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.355483 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.357031 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.357107 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.357141 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.357163 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.357176 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.459859 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.459907 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.459916 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.459936 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.459946 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.474099 4858 generic.go:334] "Generic (PLEG): container finished" podID="e28e3e9c-e621-4e85-af97-c3f48adb269d" containerID="b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e" exitCode=0 Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.474146 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerDied","Data":"b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.500037 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.514085 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.529448 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.547383 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.565759 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.565825 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.565839 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.565861 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.565876 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.568343 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.597331 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.635262 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.669911 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.669964 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.669980 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.670007 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.670025 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.675272 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.714996 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.759537 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.773325 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.773367 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.773378 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.773397 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.773411 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.796338 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.839209 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.876021 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.876076 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.876087 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.876109 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.876122 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.878171 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.921810 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:42Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.979234 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.979274 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.979285 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.979301 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:42 crc kubenswrapper[4858]: I0320 08:58:42.979335 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:42Z","lastTransitionTime":"2026-03-20T08:58:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.069256 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.069370 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:43 crc kubenswrapper[4858]: E0320 08:58:43.069470 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.069370 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:43 crc kubenswrapper[4858]: E0320 08:58:43.069611 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:43 crc kubenswrapper[4858]: E0320 08:58:43.069686 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.083120 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.083228 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.083241 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.083257 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.083268 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.186241 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.186342 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.186362 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.186397 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.186434 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.289474 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.289515 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.289526 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.289544 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.289557 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.392802 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.393241 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.393254 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.393272 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.393284 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.482969 4858 generic.go:334] "Generic (PLEG): container finished" podID="e28e3e9c-e621-4e85-af97-c3f48adb269d" containerID="8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48" exitCode=0 Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.483052 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerDied","Data":"8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.498996 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.499045 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.499061 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.499084 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.499102 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.510407 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.539241 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.562244 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.582521 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.596223 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.608814 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.608919 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.608929 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.608949 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.608961 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.611209 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.629330 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.645134 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.658915 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.674981 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.696399 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.711628 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.711716 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.711736 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.711796 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.711820 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.712492 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.729792 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.745934 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:43Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.815302 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.815356 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.815366 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.815383 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.815395 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.917294 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.917345 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.917358 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.917378 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:43 crc kubenswrapper[4858]: I0320 08:58:43.917391 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:43Z","lastTransitionTime":"2026-03-20T08:58:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.020352 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.020382 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.020391 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.020406 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.020415 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.124336 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.124392 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.124403 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.124427 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.124440 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.227525 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.227568 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.227581 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.227601 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.227614 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.331593 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.331632 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.331647 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.331667 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.331682 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.434137 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.434175 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.434185 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.434203 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.434215 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.493595 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" event={"ID":"e28e3e9c-e621-4e85-af97-c3f48adb269d","Type":"ContainerStarted","Data":"cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.505519 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.505881 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.513238 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.534859 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.537432 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.537467 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.537478 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.537496 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.537509 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.537955 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.556608 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.572223 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.588009 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.601254 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.616932 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.633557 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.639877 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.639927 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.639941 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.639961 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.639977 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.656893 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.676700 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.691901 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.716250 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.742525 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.744846 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.744912 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.744928 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.744952 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.744968 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.757192 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.771504 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.789441 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.805799 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.817913 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.830369 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.842301 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.847186 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.847239 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.847255 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.847276 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.847290 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.852915 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.867011 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.884069 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.900525 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.905008 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.905073 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.905092 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.905148 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.905165 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.916634 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: E0320 08:58:44.920084 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.924391 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.924432 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.924449 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.924473 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.924490 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.938375 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: E0320 08:58:44.938678 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.942852 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.942895 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.942907 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.942924 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.942938 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.955062 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: E0320 08:58:44.957813 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.962632 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.962685 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.962697 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.962714 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.962727 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.971146 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: E0320 08:58:44.976266 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:44Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.995154 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.995226 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.995238 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.995258 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:44 crc kubenswrapper[4858]: I0320 08:58:44.995267 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:44Z","lastTransitionTime":"2026-03-20T08:58:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: E0320 08:58:45.011337 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: E0320 08:58:45.011466 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.013924 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.013953 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.013962 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.013977 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.013987 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.069870 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.069921 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:45 crc kubenswrapper[4858]: E0320 08:58:45.070027 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.069895 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:45 crc kubenswrapper[4858]: E0320 08:58:45.070171 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:45 crc kubenswrapper[4858]: E0320 08:58:45.070251 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.116338 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.116393 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.116404 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.116424 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.116434 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.218908 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.218960 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.218973 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.218991 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.219001 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.322088 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.322162 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.322182 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.322209 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.322230 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.425821 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.425879 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.425890 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.425909 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.425923 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.510494 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.510571 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.528782 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.528833 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.528852 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.528872 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.528885 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.540486 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.565666 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.582601 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.599587 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.617263 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.631890 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.631942 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.631956 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.631974 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.631992 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.633501 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.652143 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.672261 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.687669 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.703794 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.721047 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.734977 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.735271 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.735390 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.735497 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.735565 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.738994 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.754759 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.766815 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.778495 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:45Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.838592 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.838639 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.838651 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.838670 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.838681 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.941797 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.941843 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.941853 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.941871 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:45 crc kubenswrapper[4858]: I0320 08:58:45.941883 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:45Z","lastTransitionTime":"2026-03-20T08:58:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.044633 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.044702 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.044714 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.044734 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.044747 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.071300 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.151054 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.151101 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.151112 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.151135 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.151148 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.254512 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.254560 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.254571 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.254609 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.254620 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.357496 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.357563 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.357577 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.357595 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.357606 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.459862 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.459924 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.459943 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.459971 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.459988 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.515055 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.516859 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.531413 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.544518 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.604334 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.604375 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.604386 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.604405 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.604416 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.607277 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.620703 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.632299 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.644729 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.656483 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.667091 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.678592 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.695951 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.705433 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.707191 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.707223 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.707232 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.707249 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.707260 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.720176 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.737276 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.754824 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:46Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.810679 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.810749 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.810763 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.810795 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.810811 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.913686 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.913725 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.913734 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.913749 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:46 crc kubenswrapper[4858]: I0320 08:58:46.913759 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:46Z","lastTransitionTime":"2026-03-20T08:58:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.016410 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.016456 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.016470 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.016497 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.016511 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.070036 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.070075 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.070210 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:47 crc kubenswrapper[4858]: E0320 08:58:47.070199 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:47 crc kubenswrapper[4858]: E0320 08:58:47.070405 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:47 crc kubenswrapper[4858]: E0320 08:58:47.070551 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.119475 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.119518 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.119530 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.119549 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.119560 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.222115 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.222155 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.222165 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.222184 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.222197 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.324953 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.324996 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.325007 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.325028 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.325042 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.428505 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.428557 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.428571 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.428593 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.428606 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.521551 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/0.log" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.523930 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe" exitCode=1 Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.523989 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.524624 4858 scope.go:117] "RemoveContainer" containerID="d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.530916 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.530962 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.530972 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.530989 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.531016 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.540550 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.559505 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.573675 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.586498 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.605587 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.624852 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.634389 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.634430 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.634444 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.634465 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.634478 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.637139 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.650353 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.672855 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:46Z\\\",\\\"message\\\":\\\"val\\\\nI0320 08:58:46.872852 6654 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:46.872888 6654 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:46.872889 6654 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:46.872945 6654 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:46.872946 6654 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:46.872928 6654 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:46.872966 6654 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0320 08:58:46.872981 6654 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:46.873002 6654 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:46.873027 6654 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0320 08:58:46.873068 6654 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:46.873081 6654 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:46.873109 6654 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 08:58:46.873118 6654 factory.go:656] Stopping watch factory\\\\nI0320 08:58:46.873123 6654 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:46.873139 6654 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.695772 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.711230 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.724508 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.737088 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.737127 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.737139 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.737158 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.737172 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.742295 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.754535 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:47Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.839438 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.839496 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.839509 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.839531 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.839544 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.942901 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.942945 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.942957 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.942977 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:47 crc kubenswrapper[4858]: I0320 08:58:47.942990 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:47Z","lastTransitionTime":"2026-03-20T08:58:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.045484 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.045538 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.045552 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.045571 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.045586 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.148700 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.148756 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.148766 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.148785 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.148797 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.251647 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.251695 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.251710 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.251732 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.251744 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.354085 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.354145 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.354157 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.354177 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.354191 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.456931 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.456970 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.456980 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.457002 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.457013 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.529239 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/1.log" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.530015 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/0.log" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.533953 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759" exitCode=1 Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.534017 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.534071 4858 scope.go:117] "RemoveContainer" containerID="d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.534957 4858 scope.go:117] "RemoveContainer" containerID="745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759" Mar 20 08:58:48 crc kubenswrapper[4858]: E0320 08:58:48.535185 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.553212 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.559430 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.559474 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.559488 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.559507 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.559523 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.566707 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.583496 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.597645 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.609400 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.626829 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4da3cf0d6bc7e942f69c3db715a99e2768ace4c7cdfeb1d9033497bcfb56dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:46Z\\\",\\\"message\\\":\\\"val\\\\nI0320 08:58:46.872852 6654 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:46.872888 6654 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:46.872889 6654 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:46.872945 6654 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:46.872946 6654 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:46.872928 6654 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:46.872966 6654 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0320 08:58:46.872981 6654 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:46.873002 6654 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:46.873027 6654 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0320 08:58:46.873068 6654 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:46.873081 6654 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:46.873109 6654 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0320 08:58:46.873118 6654 factory.go:656] Stopping watch factory\\\\nI0320 08:58:46.873123 6654 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:46.873139 6654 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.646871 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.658617 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.662227 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.662266 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.662276 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.662295 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.662306 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.674361 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.688533 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.701806 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.711276 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.721299 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.731080 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:48Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.765058 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.765093 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.765102 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.765122 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.765132 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.868384 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.868420 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.868430 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.868447 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.868457 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.971199 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.971253 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.971270 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.971299 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:48 crc kubenswrapper[4858]: I0320 08:58:48.971354 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:48Z","lastTransitionTime":"2026-03-20T08:58:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.069047 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:49 crc kubenswrapper[4858]: E0320 08:58:49.069622 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.069996 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:49 crc kubenswrapper[4858]: E0320 08:58:49.070087 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.070153 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:49 crc kubenswrapper[4858]: E0320 08:58:49.070213 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.075168 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.075444 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.075606 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.075718 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.075802 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.179535 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.179864 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.179960 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.180051 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.180147 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.282880 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.282928 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.282939 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.282959 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.282970 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.385440 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.385501 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.385519 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.385540 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.385555 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.489160 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.489196 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.489204 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.489219 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.489230 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.539414 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/1.log" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.542789 4858 scope.go:117] "RemoveContainer" containerID="745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759" Mar 20 08:58:49 crc kubenswrapper[4858]: E0320 08:58:49.542972 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.555861 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.568226 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.578063 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.591811 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.591858 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.591867 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.591886 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.591895 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.599257 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.609740 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.623586 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.637793 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.654061 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.665599 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.679608 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.690523 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.694142 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.694187 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.694199 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.694220 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.694231 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.702884 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.714804 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.728349 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.796744 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.796792 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.796804 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.796824 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.796835 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.899765 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.899803 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.899813 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.899831 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.899842 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:49Z","lastTransitionTime":"2026-03-20T08:58:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.908082 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l"] Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.909277 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.913607 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.914533 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.929735 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.944281 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75db206b-1ea7-4295-85ae-10309c438903-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.944369 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvzmh\" (UniqueName: \"kubernetes.io/projected/75db206b-1ea7-4295-85ae-10309c438903-kube-api-access-hvzmh\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.944412 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75db206b-1ea7-4295-85ae-10309c438903-env-overrides\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.944433 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75db206b-1ea7-4295-85ae-10309c438903-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.946925 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.962417 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.972193 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.983945 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:49 crc kubenswrapper[4858]: I0320 08:58:49.995515 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:49Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.003475 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.003544 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.003561 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.003590 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.003608 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.008474 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.020925 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.033983 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.045466 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75db206b-1ea7-4295-85ae-10309c438903-env-overrides\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.045514 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75db206b-1ea7-4295-85ae-10309c438903-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.045552 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75db206b-1ea7-4295-85ae-10309c438903-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.045612 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvzmh\" (UniqueName: \"kubernetes.io/projected/75db206b-1ea7-4295-85ae-10309c438903-kube-api-access-hvzmh\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.046496 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/75db206b-1ea7-4295-85ae-10309c438903-env-overrides\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.046857 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/75db206b-1ea7-4295-85ae-10309c438903-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.053854 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.054277 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/75db206b-1ea7-4295-85ae-10309c438903-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.066414 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvzmh\" (UniqueName: \"kubernetes.io/projected/75db206b-1ea7-4295-85ae-10309c438903-kube-api-access-hvzmh\") pod \"ovnkube-control-plane-749d76644c-22d5l\" (UID: \"75db206b-1ea7-4295-85ae-10309c438903\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.071504 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.086910 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.107206 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.107503 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.107595 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.107687 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.107769 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.108037 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.124147 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.138749 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.154347 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.169461 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.192569 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.203650 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.210035 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.210091 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.210102 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.210123 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.210134 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.216385 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.229382 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.230483 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.242659 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.258634 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.280383 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.290686 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.303456 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.315411 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.315749 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.315789 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.315808 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.315818 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.316640 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.328887 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.341211 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.353098 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.418622 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.418653 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.418662 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.418679 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.418689 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.521750 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.521797 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.521808 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.521828 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.521843 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.551480 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" event={"ID":"75db206b-1ea7-4295-85ae-10309c438903","Type":"ContainerStarted","Data":"4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.551541 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" event={"ID":"75db206b-1ea7-4295-85ae-10309c438903","Type":"ContainerStarted","Data":"9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.551552 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" event={"ID":"75db206b-1ea7-4295-85ae-10309c438903","Type":"ContainerStarted","Data":"1474bf32fc0b8ca8a7ba944b969f5e9ffb7b9b48ace9fdbddf183181a5d011f6"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.568938 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.589083 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.608719 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.623894 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.624582 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.624637 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.624650 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.624671 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.624686 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.636969 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.648188 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.657740 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.670048 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.685840 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.697977 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.709869 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.728086 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.728141 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.728152 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.728174 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.728192 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.728743 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.738837 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.751241 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.764690 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:50Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.830944 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.830988 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.830997 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.831013 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.831023 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.933533 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.933603 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.933614 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.933632 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:50 crc kubenswrapper[4858]: I0320 08:58:50.933642 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:50Z","lastTransitionTime":"2026-03-20T08:58:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.037131 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.037178 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.037190 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.037207 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.037218 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.069206 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.069337 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:51 crc kubenswrapper[4858]: E0320 08:58:51.069419 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.069307 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:51 crc kubenswrapper[4858]: E0320 08:58:51.069534 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:51 crc kubenswrapper[4858]: E0320 08:58:51.069634 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.140529 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.140829 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.140840 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.140862 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.140876 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.244258 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.244378 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.244395 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.244418 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.244440 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.347518 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.347560 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.347570 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.347586 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.347600 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.450618 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.450667 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.450678 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.450698 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.450714 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.552841 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.552873 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.552881 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.552896 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.552929 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.655780 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.655852 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.655874 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.655906 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.655930 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.740602 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-kvlch"] Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.741262 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:51 crc kubenswrapper[4858]: E0320 08:58:51.741368 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.757555 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.758299 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.758347 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.758358 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.758374 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.758384 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.758902 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk4cj\" (UniqueName: \"kubernetes.io/projected/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-kube-api-access-rk4cj\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.759099 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.782576 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.802268 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.823669 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.838363 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.848003 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.858136 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.859631 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.859687 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk4cj\" (UniqueName: \"kubernetes.io/projected/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-kube-api-access-rk4cj\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:51 crc kubenswrapper[4858]: E0320 08:58:51.859827 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:51 crc kubenswrapper[4858]: E0320 08:58:51.859963 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 08:58:52.359933785 +0000 UTC m=+113.680351982 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.861373 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.861413 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.861425 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.861443 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.861453 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.871906 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.877671 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk4cj\" (UniqueName: \"kubernetes.io/projected/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-kube-api-access-rk4cj\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.882368 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.894580 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.911757 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.927713 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.940237 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.952971 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.963868 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.964070 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.964129 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.964210 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.964296 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:51Z","lastTransitionTime":"2026-03-20T08:58:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.968064 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:51 crc kubenswrapper[4858]: I0320 08:58:51.981879 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:51Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.067302 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.067363 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.067375 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.067397 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.067410 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.170023 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.170444 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.170514 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.170609 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.170701 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.273957 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.274101 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.274178 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.274335 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.274441 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.366082 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:52 crc kubenswrapper[4858]: E0320 08:58:52.366346 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:52 crc kubenswrapper[4858]: E0320 08:58:52.367155 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 08:58:53.367093693 +0000 UTC m=+114.687511890 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.377169 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.377240 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.377258 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.377287 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.377306 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.480780 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.480969 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.481037 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.481117 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.481174 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.583496 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.583935 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.584000 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.584065 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.584124 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.687604 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.688207 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.688540 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.688731 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.688820 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.792141 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.792184 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.792194 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.792210 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.792221 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.895515 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.895554 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.895566 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.895585 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.895595 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.998432 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.998484 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.998497 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.998519 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:52 crc kubenswrapper[4858]: I0320 08:58:52.998531 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:52Z","lastTransitionTime":"2026-03-20T08:58:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.070004 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.070006 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.070073 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:53 crc kubenswrapper[4858]: E0320 08:58:53.070552 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:53 crc kubenswrapper[4858]: E0320 08:58:53.070652 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:53 crc kubenswrapper[4858]: E0320 08:58:53.070600 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.101681 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.101988 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.102094 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.102171 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.102236 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.205527 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.205782 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.205956 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.206059 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.206234 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.309498 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.309589 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.309612 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.309643 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.309660 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.377484 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:53 crc kubenswrapper[4858]: E0320 08:58:53.377804 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:53 crc kubenswrapper[4858]: E0320 08:58:53.378008 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 08:58:55.377967565 +0000 UTC m=+116.698385822 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.413121 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.413177 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.413189 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.413209 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.413223 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.516295 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.516601 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.516687 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.516788 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.516913 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.619943 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.619980 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.620007 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.620047 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.620061 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.723085 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.723130 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.723144 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.723161 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.723172 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.826354 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.826618 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.826724 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.826806 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.826876 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.930630 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.931083 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.931172 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.931268 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:53 crc kubenswrapper[4858]: I0320 08:58:53.931386 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:53Z","lastTransitionTime":"2026-03-20T08:58:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.029967 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.034673 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.034721 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.034734 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.034775 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.034793 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.069954 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.070130 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.137132 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.137451 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.137481 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.137509 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.137520 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.240258 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.240348 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.240366 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.240393 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.240411 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.344702 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.344771 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.344783 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.344805 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.344818 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.447278 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.447376 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.447396 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.447421 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.447439 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.550580 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.550649 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.550659 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.550679 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.550692 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.653488 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.653538 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.653551 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.653572 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.653584 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.758414 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.758470 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.758484 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.758506 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.758519 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.861557 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.861603 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.861614 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.861629 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.861640 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.897937 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.898241 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 08:59:26.898205207 +0000 UTC m=+148.218623404 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.965017 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.965075 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.965101 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.965123 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.965135 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:54Z","lastTransitionTime":"2026-03-20T08:58:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.999142 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999428 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999456 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999472 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.999570 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999720 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:26.999634806 +0000 UTC m=+148.320053003 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999763 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999802 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999819 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.999822 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:54 crc kubenswrapper[4858]: E0320 08:58:54.999898 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:26.999871303 +0000 UTC m=+148.320289700 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:58:54 crc kubenswrapper[4858]: I0320 08:58:54.999943 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.000046 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.000064 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.000110 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:27.000102739 +0000 UTC m=+148.320521166 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.000131 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 08:59:27.00012278 +0000 UTC m=+148.320541197 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069046 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069233 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069612 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069712 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069743 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069352 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069279 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.069932 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.070073 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.070151 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.069846 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.177794 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.177855 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.177871 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.177897 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.177917 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.192750 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.196482 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.196534 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.196545 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.196566 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.196577 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.209091 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.213232 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.213271 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.213280 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.213300 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.213329 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.224439 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.228097 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.228153 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.228167 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.228191 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.228204 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.240286 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.244183 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.244234 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.244249 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.244276 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.244292 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.257376 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:58:55Z is after 2025-08-24T17:21:41Z" Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.257507 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.259733 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.259774 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.259788 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.259808 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.259821 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.362584 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.362641 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.362658 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.362687 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.362706 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.405234 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.405414 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:55 crc kubenswrapper[4858]: E0320 08:58:55.405492 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 08:58:59.405472568 +0000 UTC m=+120.725890765 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.465597 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.465895 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.465965 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.466039 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.466097 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.569653 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.570408 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.570636 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.570836 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.571015 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.673793 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.673991 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.674008 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.674039 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.674049 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.776425 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.776474 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.776487 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.776513 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.776529 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.879475 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.879528 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.879541 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.879564 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.879578 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.982816 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.982864 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.982884 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.982901 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:55 crc kubenswrapper[4858]: I0320 08:58:55.982915 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:55Z","lastTransitionTime":"2026-03-20T08:58:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.069735 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:56 crc kubenswrapper[4858]: E0320 08:58:56.069967 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.085896 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.085956 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.085971 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.085991 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.086007 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.189189 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.189242 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.189253 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.189271 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.189283 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.292772 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.292829 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.292843 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.292869 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.292886 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.395735 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.395787 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.395799 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.395829 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.395841 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.498952 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.499006 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.499019 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.499037 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.499051 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.601589 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.601638 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.601650 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.601671 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.601683 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.705002 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.705055 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.705067 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.705089 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.705101 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.808410 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.808459 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.808468 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.808487 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.808502 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.912304 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.912384 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.912398 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.912423 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:56 crc kubenswrapper[4858]: I0320 08:58:56.912436 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:56Z","lastTransitionTime":"2026-03-20T08:58:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.015148 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.015189 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.015198 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.015215 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.015225 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.070074 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.070110 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.070074 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:57 crc kubenswrapper[4858]: E0320 08:58:57.070238 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:57 crc kubenswrapper[4858]: E0320 08:58:57.070398 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:57 crc kubenswrapper[4858]: E0320 08:58:57.070521 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.117794 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.117851 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.117865 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.117885 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.117897 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.219948 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.219990 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.220002 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.220019 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.220029 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.322956 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.323012 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.323021 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.323041 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.323054 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.427747 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.427826 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.427839 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.427863 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.427877 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.531267 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.531335 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.531348 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.531371 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.531384 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.634192 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.634243 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.634256 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.634279 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.634296 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.737276 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.737340 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.737358 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.737382 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.737393 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.839647 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.839699 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.839711 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.839733 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.839746 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.942189 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.942267 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.942278 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.942298 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:57 crc kubenswrapper[4858]: I0320 08:58:57.942337 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:57Z","lastTransitionTime":"2026-03-20T08:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.044884 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.044935 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.044952 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.044974 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.044989 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.069268 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:58 crc kubenswrapper[4858]: E0320 08:58:58.069465 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.147883 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.147930 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.147940 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.147958 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.147969 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.251460 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.251511 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.251520 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.251541 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.251550 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.356506 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.356588 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.356604 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.356631 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.356653 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.459468 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.459525 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.459538 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.459558 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.459571 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.562867 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.562908 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.562917 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.562938 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.562949 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.666425 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.666512 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.666527 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.666546 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.666556 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.769682 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.769779 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.769800 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.769828 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.769846 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.872720 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.872784 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.872796 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.872817 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.872833 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.976175 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.976220 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.976229 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.976249 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:58 crc kubenswrapper[4858]: I0320 08:58:58.976260 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:58Z","lastTransitionTime":"2026-03-20T08:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.069189 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.069394 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.069469 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:58:59 crc kubenswrapper[4858]: E0320 08:58:59.069441 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:58:59 crc kubenswrapper[4858]: E0320 08:58:59.069756 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:58:59 crc kubenswrapper[4858]: E0320 08:58:59.069830 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.078803 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.078957 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.078977 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.078999 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.079011 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.182088 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.182197 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.182271 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.182380 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.182466 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.285858 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.285903 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.285913 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.285931 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.285942 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.423557 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.423639 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.423652 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.423675 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.423689 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.455614 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:58:59 crc kubenswrapper[4858]: E0320 08:58:59.455819 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:59 crc kubenswrapper[4858]: E0320 08:58:59.455908 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 08:59:07.455878627 +0000 UTC m=+128.776296824 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.526704 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.526747 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.526756 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.526773 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.526782 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.629905 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.629959 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.629971 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.629992 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.630003 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.733069 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.733120 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.733130 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.733152 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.733172 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.836232 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.836342 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.836369 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.836398 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.836421 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.939861 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.939901 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.939910 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.939929 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:58:59 crc kubenswrapper[4858]: I0320 08:58:59.939940 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:58:59Z","lastTransitionTime":"2026-03-20T08:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:00 crc kubenswrapper[4858]: E0320 08:59:00.041018 4858 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.069907 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:00 crc kubenswrapper[4858]: E0320 08:59:00.070137 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.086497 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.103492 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.116439 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.126688 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.136987 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.147864 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.160188 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.173419 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: E0320 08:59:00.181381 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.189193 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.204186 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.220233 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.239658 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.251254 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.265971 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.278586 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:00 crc kubenswrapper[4858]: I0320 08:59:00.289918 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:00Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:01 crc kubenswrapper[4858]: I0320 08:59:01.069438 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:01 crc kubenswrapper[4858]: I0320 08:59:01.069569 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:01 crc kubenswrapper[4858]: I0320 08:59:01.069811 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:01 crc kubenswrapper[4858]: E0320 08:59:01.069792 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:01 crc kubenswrapper[4858]: E0320 08:59:01.069866 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:01 crc kubenswrapper[4858]: E0320 08:59:01.069600 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.069616 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:02 crc kubenswrapper[4858]: E0320 08:59:02.069780 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.070443 4858 scope.go:117] "RemoveContainer" containerID="745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.593999 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/1.log" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.597047 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9"} Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.597529 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.617042 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.629678 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.643055 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.659534 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.674082 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.687232 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.703345 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.718184 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.732927 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.750551 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.769845 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.783591 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.796483 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.808227 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.821204 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:02 crc kubenswrapper[4858]: I0320 08:59:02.835608 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:02Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.069558 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:03 crc kubenswrapper[4858]: E0320 08:59:03.069990 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.069666 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:03 crc kubenswrapper[4858]: E0320 08:59:03.071245 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.069558 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:03 crc kubenswrapper[4858]: E0320 08:59:03.071610 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.603227 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/2.log" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.604006 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/1.log" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.607617 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9" exitCode=1 Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.607705 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9"} Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.607753 4858 scope.go:117] "RemoveContainer" containerID="745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.609088 4858 scope.go:117] "RemoveContainer" containerID="c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9" Mar 20 08:59:03 crc kubenswrapper[4858]: E0320 08:59:03.609298 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.620172 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.631859 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.643648 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.654263 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.669011 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.692376 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.709067 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.722540 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.734454 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.748977 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.768034 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.778974 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.790907 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.801406 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.812115 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:03 crc kubenswrapper[4858]: I0320 08:59:03.823577 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:03Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.033729 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.047388 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.068067 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://745966f984424e426f576c559249da98cd08dc2836231ea287f972316ca45759\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:58:48Z\\\",\\\"message\\\":\\\"0 08:58:48.307314 6820 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0320 08:58:48.307392 6820 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0320 08:58:48.307428 6820 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0320 08:58:48.308725 6820 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0320 08:58:48.308761 6820 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0320 08:58:48.308783 6820 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0320 08:58:48.308787 6820 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0320 08:58:48.308792 6820 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0320 08:58:48.308808 6820 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0320 08:58:48.308818 6820 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0320 08:58:48.308830 6820 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0320 08:58:48.308838 6820 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0320 08:58:48.308859 6820 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0320 08:58:48.308874 6820 handler.go:208] Removed *v1.Node event handler 2\\\\nI0320 08:58:48.308881 6820 factory.go:656] Stopping watch factory\\\\nI0320 08:58:48.308897 6820 ovnkube.go:599] Stopped ovnkube\\\\nI03\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.069404 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:04 crc kubenswrapper[4858]: E0320 08:59:04.069535 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.077939 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.079990 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.090495 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.104238 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.114739 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.128188 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.141401 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.153958 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.167198 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.179075 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.189396 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.198858 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.208802 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.218201 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.227600 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.614818 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/2.log" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.620263 4858 scope.go:117] "RemoveContainer" containerID="c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9" Mar 20 08:59:04 crc kubenswrapper[4858]: E0320 08:59:04.620512 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.636966 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.656554 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.681383 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.694027 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.708487 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.723044 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.738623 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.754429 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.767037 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.781086 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.793977 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.805616 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.817010 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.826497 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.835728 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.847211 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:04 crc kubenswrapper[4858]: I0320 08:59:04.859127 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:04Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.069464 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.069531 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.069650 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.069789 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.069944 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.070202 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.182581 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.447807 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.447867 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.447879 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.447903 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.447916 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.461423 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.465978 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.466097 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.466387 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.466417 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.466467 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.479088 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.482699 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.482736 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.482780 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.482801 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.482826 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.494955 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.498414 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.498534 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.498617 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.498685 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.498744 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.510744 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.514191 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.514240 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.514252 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.514271 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:05 crc kubenswrapper[4858]: I0320 08:59:05.514284 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:05Z","lastTransitionTime":"2026-03-20T08:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.528193 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:05Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:05 crc kubenswrapper[4858]: E0320 08:59:05.528528 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:06 crc kubenswrapper[4858]: I0320 08:59:06.069362 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:06 crc kubenswrapper[4858]: E0320 08:59:06.069555 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:06 crc kubenswrapper[4858]: I0320 08:59:06.079679 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 20 08:59:06 crc kubenswrapper[4858]: I0320 08:59:06.487993 4858 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 08:59:07 crc kubenswrapper[4858]: I0320 08:59:07.070022 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:07 crc kubenswrapper[4858]: I0320 08:59:07.070105 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:07 crc kubenswrapper[4858]: I0320 08:59:07.070164 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:07 crc kubenswrapper[4858]: E0320 08:59:07.070283 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:07 crc kubenswrapper[4858]: E0320 08:59:07.070437 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:07 crc kubenswrapper[4858]: E0320 08:59:07.070661 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:07 crc kubenswrapper[4858]: I0320 08:59:07.546071 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:07 crc kubenswrapper[4858]: E0320 08:59:07.546357 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:07 crc kubenswrapper[4858]: E0320 08:59:07.546495 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 08:59:23.54646245 +0000 UTC m=+144.866880677 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:08 crc kubenswrapper[4858]: I0320 08:59:08.069454 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:08 crc kubenswrapper[4858]: E0320 08:59:08.069990 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:09 crc kubenswrapper[4858]: I0320 08:59:09.069666 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:09 crc kubenswrapper[4858]: I0320 08:59:09.069820 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:09 crc kubenswrapper[4858]: I0320 08:59:09.069666 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:09 crc kubenswrapper[4858]: E0320 08:59:09.069902 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:09 crc kubenswrapper[4858]: E0320 08:59:09.070013 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:09 crc kubenswrapper[4858]: E0320 08:59:09.070134 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.069413 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:10 crc kubenswrapper[4858]: E0320 08:59:10.069604 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.084165 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.098221 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.112735 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.126939 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.142751 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.155970 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.168744 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: E0320 08:59:10.183479 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.192368 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.219093 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.236342 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.257889 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.277053 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.291517 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.308980 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.321664 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.338134 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.352822 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:10 crc kubenswrapper[4858]: I0320 08:59:10.368562 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:10Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:11 crc kubenswrapper[4858]: I0320 08:59:11.069479 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:11 crc kubenswrapper[4858]: I0320 08:59:11.069498 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:11 crc kubenswrapper[4858]: I0320 08:59:11.069656 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:11 crc kubenswrapper[4858]: E0320 08:59:11.069954 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:11 crc kubenswrapper[4858]: E0320 08:59:11.070062 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:11 crc kubenswrapper[4858]: E0320 08:59:11.070132 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:12 crc kubenswrapper[4858]: I0320 08:59:12.069609 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:12 crc kubenswrapper[4858]: E0320 08:59:12.070376 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:13 crc kubenswrapper[4858]: I0320 08:59:13.069166 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:13 crc kubenswrapper[4858]: I0320 08:59:13.069241 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:13 crc kubenswrapper[4858]: E0320 08:59:13.069345 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:13 crc kubenswrapper[4858]: I0320 08:59:13.069418 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:13 crc kubenswrapper[4858]: E0320 08:59:13.069478 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:13 crc kubenswrapper[4858]: E0320 08:59:13.069616 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:14 crc kubenswrapper[4858]: I0320 08:59:14.069700 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:14 crc kubenswrapper[4858]: E0320 08:59:14.069861 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.069760 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.069808 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.069830 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.069948 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.070063 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.070196 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.184651 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.726123 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.726183 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.726197 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.726217 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.726232 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:15Z","lastTransitionTime":"2026-03-20T08:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.749332 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:15Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.753205 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.753288 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.753366 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.753414 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.753447 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:15Z","lastTransitionTime":"2026-03-20T08:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.766881 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:15Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.770903 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.770961 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.770973 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.770995 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.771007 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:15Z","lastTransitionTime":"2026-03-20T08:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.783351 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:15Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.787676 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.787712 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.787723 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.787741 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.787752 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:15Z","lastTransitionTime":"2026-03-20T08:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.800184 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:15Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.803747 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.803778 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.803788 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.803804 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:15 crc kubenswrapper[4858]: I0320 08:59:15.803816 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:15Z","lastTransitionTime":"2026-03-20T08:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.819056 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:15Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:15 crc kubenswrapper[4858]: E0320 08:59:15.819258 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:16 crc kubenswrapper[4858]: I0320 08:59:16.069690 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:16 crc kubenswrapper[4858]: E0320 08:59:16.069869 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:16 crc kubenswrapper[4858]: I0320 08:59:16.070704 4858 scope.go:117] "RemoveContainer" containerID="c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9" Mar 20 08:59:16 crc kubenswrapper[4858]: E0320 08:59:16.071012 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:59:17 crc kubenswrapper[4858]: I0320 08:59:17.069821 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:17 crc kubenswrapper[4858]: I0320 08:59:17.069864 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:17 crc kubenswrapper[4858]: I0320 08:59:17.069861 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:17 crc kubenswrapper[4858]: E0320 08:59:17.069989 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:17 crc kubenswrapper[4858]: E0320 08:59:17.070165 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:17 crc kubenswrapper[4858]: E0320 08:59:17.070283 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:18 crc kubenswrapper[4858]: I0320 08:59:18.069699 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:18 crc kubenswrapper[4858]: E0320 08:59:18.069937 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:19 crc kubenswrapper[4858]: I0320 08:59:19.070137 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:19 crc kubenswrapper[4858]: E0320 08:59:19.070558 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:19 crc kubenswrapper[4858]: I0320 08:59:19.070239 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:19 crc kubenswrapper[4858]: E0320 08:59:19.071041 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:19 crc kubenswrapper[4858]: I0320 08:59:19.070205 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:19 crc kubenswrapper[4858]: E0320 08:59:19.071215 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:19 crc kubenswrapper[4858]: I0320 08:59:19.692847 4858 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.069557 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:20 crc kubenswrapper[4858]: E0320 08:59:20.069773 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.083178 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.094642 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.108118 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.121529 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.132141 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.144277 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.159152 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.170287 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: E0320 08:59:20.185447 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.194406 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.221743 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.237612 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.257704 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.275356 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.289796 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.305608 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.323414 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.339539 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:20 crc kubenswrapper[4858]: I0320 08:59:20.353261 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:20Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:21 crc kubenswrapper[4858]: I0320 08:59:21.069450 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:21 crc kubenswrapper[4858]: I0320 08:59:21.069555 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:21 crc kubenswrapper[4858]: E0320 08:59:21.069659 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:21 crc kubenswrapper[4858]: I0320 08:59:21.069480 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:21 crc kubenswrapper[4858]: E0320 08:59:21.069873 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:21 crc kubenswrapper[4858]: E0320 08:59:21.069970 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:22 crc kubenswrapper[4858]: I0320 08:59:22.070138 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:22 crc kubenswrapper[4858]: E0320 08:59:22.070362 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:23 crc kubenswrapper[4858]: I0320 08:59:23.069298 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:23 crc kubenswrapper[4858]: I0320 08:59:23.069304 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:23 crc kubenswrapper[4858]: E0320 08:59:23.069503 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:23 crc kubenswrapper[4858]: E0320 08:59:23.069502 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:23 crc kubenswrapper[4858]: I0320 08:59:23.069341 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:23 crc kubenswrapper[4858]: E0320 08:59:23.069608 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:23 crc kubenswrapper[4858]: I0320 08:59:23.628813 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:23 crc kubenswrapper[4858]: E0320 08:59:23.629078 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:23 crc kubenswrapper[4858]: E0320 08:59:23.629199 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 08:59:55.629165557 +0000 UTC m=+176.949583944 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.068933 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:24 crc kubenswrapper[4858]: E0320 08:59:24.069092 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.694738 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/0.log" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.694804 4858 generic.go:334] "Generic (PLEG): container finished" podID="24656c62-314b-4c20-adf1-217d58a95f57" containerID="574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0" exitCode=1 Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.694848 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerDied","Data":"574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0"} Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.695411 4858 scope.go:117] "RemoveContainer" containerID="574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.720242 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.733275 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.747234 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.761734 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.785891 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.798365 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.815746 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.833910 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.848205 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.861021 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.874886 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.892820 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6\\\\n2026-03-20T08:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:39Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:39Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.907692 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.919733 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.929364 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.938914 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.950515 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:24 crc kubenswrapper[4858]: I0320 08:59:24.961219 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:24Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.069336 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.069387 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.069345 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.069486 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.069575 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.069676 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.186603 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.700669 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/0.log" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.700746 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerStarted","Data":"3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020"} Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.717226 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.731155 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.752635 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.765876 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6\\\\n2026-03-20T08:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:39Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:39Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.780141 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.793484 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.805993 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.817104 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.827625 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.831213 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.831256 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.831271 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.831298 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.831337 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.839093 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.843652 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.846770 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.846815 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.846825 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.846845 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.846856 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.852029 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.859591 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.863267 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.863301 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.863311 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.863347 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.863359 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.868134 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.875041 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.878427 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.878471 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.878483 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.878504 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.878522 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.881297 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.890329 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.893862 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.893900 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.893915 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.893939 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.893950 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:25Z","lastTransitionTime":"2026-03-20T08:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.900778 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.905227 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: E0320 08:59:25.905356 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.909833 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.920754 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.935652 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:25 crc kubenswrapper[4858]: I0320 08:59:25.949273 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:25Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:26 crc kubenswrapper[4858]: I0320 08:59:26.069920 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:26 crc kubenswrapper[4858]: E0320 08:59:26.070111 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:26 crc kubenswrapper[4858]: I0320 08:59:26.968121 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 08:59:26 crc kubenswrapper[4858]: E0320 08:59:26.968424 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.968377174 +0000 UTC m=+212.288795371 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.069749 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.069909 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.069956 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.070074 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.070282 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.070383 4858 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.070495 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.070468377 +0000 UTC m=+212.390886574 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.070107 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.070722 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.070873 4858 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.070942 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.070918453 +0000 UTC m=+212.391336850 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.070976 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.071166 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071346 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071370 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071383 4858 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071426 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.071416359 +0000 UTC m=+212.391834566 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.071578 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071684 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071704 4858 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071714 4858 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:27 crc kubenswrapper[4858]: E0320 08:59:27.071750 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.0717394 +0000 UTC m=+212.392157597 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.071944 4858 scope.go:117] "RemoveContainer" containerID="c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.711954 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/2.log" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.715447 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1"} Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.716005 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.732172 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.746063 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.759953 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6\\\\n2026-03-20T08:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:39Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:39Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.775258 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.793122 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.811365 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.827885 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.842453 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.854044 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.868947 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.883015 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.895870 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.918589 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.928606 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.941003 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.956444 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.970957 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:27 crc kubenswrapper[4858]: I0320 08:59:27.983715 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:27Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.069557 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:28 crc kubenswrapper[4858]: E0320 08:59:28.069956 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.721914 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/3.log" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.722488 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/2.log" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.725207 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" exitCode=1 Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.725261 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1"} Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.725457 4858 scope.go:117] "RemoveContainer" containerID="c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.726337 4858 scope.go:117] "RemoveContainer" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" Mar 20 08:59:28 crc kubenswrapper[4858]: E0320 08:59:28.726594 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.748710 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.764214 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.786519 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c8e80ba46e3075c07aaf8b2fcb2415eb09c990bcf3645ecbaeaf1efc445512c9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:02Z\\\",\\\"message\\\":\\\"s.io/client-go/informers/factory.go:160\\\\nI0320 08:59:02.917874 7075 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918646 7075 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.918641 7075 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0320 08:59:02.918694 7075 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0320 08:59:02.919439 7075 factory.go:656] Stopping watch factory\\\\nI0320 08:59:02.944197 7075 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI0320 08:59:02.944221 7075 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI0320 08:59:02.944268 7075 ovnkube.go:599] Stopped ovnkube\\\\nI0320 08:59:02.944291 7075 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0320 08:59:02.944391 7075 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:02Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:27Z\\\",\\\"message\\\":\\\"ces.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:27.867092 7389 services_controller.go:360] Finished syncing service packageserver-service on namespace openshift-operator-lifecycle-manager for network=default : 1.018625ms\\\\nI0320 08:59:27.867561 7389 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867575 7389 services_controller.go:453] Built service openshift-kube-controller-manager/kube-controller-manager template LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867585 7389 services_controller.go:454] Service openshift-kube-controller-manager/kube-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0320 08:59:27.867590 7389 services_controller.go:356] Processing sync for service openshift-machine-api/machine-api-operator-machine-webhook for network=default\\\\nF0320 08:59:27.867590 7389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.802409 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.813820 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.825922 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.840047 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.852956 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.869822 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.888878 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.907474 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6\\\\n2026-03-20T08:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:39Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:39Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.918783 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.932506 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.945581 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.958509 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.971361 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.983804 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:28 crc kubenswrapper[4858]: I0320 08:59:28.995738 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:28Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.069053 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.069105 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.069169 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:29 crc kubenswrapper[4858]: E0320 08:59:29.069228 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:29 crc kubenswrapper[4858]: E0320 08:59:29.069340 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:29 crc kubenswrapper[4858]: E0320 08:59:29.069463 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.730389 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/3.log" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.734401 4858 scope.go:117] "RemoveContainer" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" Mar 20 08:59:29 crc kubenswrapper[4858]: E0320 08:59:29.734611 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.747125 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.761674 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.774610 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6\\\\n2026-03-20T08:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:39Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:39Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.787263 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.800182 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.811994 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.822475 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.832143 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.843242 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.853981 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.866435 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.879584 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.899891 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:27Z\\\",\\\"message\\\":\\\"ces.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:27.867092 7389 services_controller.go:360] Finished syncing service packageserver-service on namespace openshift-operator-lifecycle-manager for network=default : 1.018625ms\\\\nI0320 08:59:27.867561 7389 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867575 7389 services_controller.go:453] Built service openshift-kube-controller-manager/kube-controller-manager template LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867585 7389 services_controller.go:454] Service openshift-kube-controller-manager/kube-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0320 08:59:27.867590 7389 services_controller.go:356] Processing sync for service openshift-machine-api/machine-api-operator-machine-webhook for network=default\\\\nF0320 08:59:27.867590 7389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.911586 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.924511 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.941227 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.956207 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:29 crc kubenswrapper[4858]: I0320 08:59:29.967853 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:29Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.069707 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:30 crc kubenswrapper[4858]: E0320 08:59:30.069923 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.083568 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.096804 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.111137 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.125636 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.141245 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6\\\\n2026-03-20T08:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:39Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:39Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.152339 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.163848 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.173428 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.187056 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: E0320 08:59:30.187901 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.199736 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.211438 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.223665 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.236469 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.249234 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.262048 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.289342 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:27Z\\\",\\\"message\\\":\\\"ces.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:27.867092 7389 services_controller.go:360] Finished syncing service packageserver-service on namespace openshift-operator-lifecycle-manager for network=default : 1.018625ms\\\\nI0320 08:59:27.867561 7389 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867575 7389 services_controller.go:453] Built service openshift-kube-controller-manager/kube-controller-manager template LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867585 7389 services_controller.go:454] Service openshift-kube-controller-manager/kube-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0320 08:59:27.867590 7389 services_controller.go:356] Processing sync for service openshift-machine-api/machine-api-operator-machine-webhook for network=default\\\\nF0320 08:59:27.867590 7389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.305757 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:30 crc kubenswrapper[4858]: I0320 08:59:30.320430 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:30Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:31 crc kubenswrapper[4858]: I0320 08:59:31.069795 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:31 crc kubenswrapper[4858]: I0320 08:59:31.069878 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:31 crc kubenswrapper[4858]: I0320 08:59:31.070602 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:31 crc kubenswrapper[4858]: E0320 08:59:31.070791 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:31 crc kubenswrapper[4858]: E0320 08:59:31.070945 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:31 crc kubenswrapper[4858]: E0320 08:59:31.071029 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:32 crc kubenswrapper[4858]: I0320 08:59:32.071042 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:32 crc kubenswrapper[4858]: E0320 08:59:32.071276 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:33 crc kubenswrapper[4858]: I0320 08:59:33.069831 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:33 crc kubenswrapper[4858]: I0320 08:59:33.069878 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:33 crc kubenswrapper[4858]: I0320 08:59:33.069834 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:33 crc kubenswrapper[4858]: E0320 08:59:33.070026 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:33 crc kubenswrapper[4858]: E0320 08:59:33.070126 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:33 crc kubenswrapper[4858]: E0320 08:59:33.070283 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:34 crc kubenswrapper[4858]: I0320 08:59:34.069999 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:34 crc kubenswrapper[4858]: E0320 08:59:34.070307 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.070009 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.070066 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.070145 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:35 crc kubenswrapper[4858]: E0320 08:59:35.070200 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:35 crc kubenswrapper[4858]: E0320 08:59:35.070382 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:35 crc kubenswrapper[4858]: E0320 08:59:35.070471 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:35 crc kubenswrapper[4858]: E0320 08:59:35.189192 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.998759 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.998812 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.998824 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.998840 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:35 crc kubenswrapper[4858]: I0320 08:59:35.998853 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:35Z","lastTransitionTime":"2026-03-20T08:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:36 crc kubenswrapper[4858]: E0320 08:59:36.013558 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.019669 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.019726 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.019753 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.019773 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.019786 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:36Z","lastTransitionTime":"2026-03-20T08:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:36 crc kubenswrapper[4858]: E0320 08:59:36.035536 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.040824 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.040909 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.040921 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.040945 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.040961 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:36Z","lastTransitionTime":"2026-03-20T08:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:36 crc kubenswrapper[4858]: E0320 08:59:36.055671 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.060044 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.060101 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.060112 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.060134 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.060150 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:36Z","lastTransitionTime":"2026-03-20T08:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.070111 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:36 crc kubenswrapper[4858]: E0320 08:59:36.070360 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:36 crc kubenswrapper[4858]: E0320 08:59:36.074861 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.078667 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.078716 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.078730 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.078749 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:36 crc kubenswrapper[4858]: I0320 08:59:36.078764 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:36Z","lastTransitionTime":"2026-03-20T08:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:36 crc kubenswrapper[4858]: E0320 08:59:36.093774 4858 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"c2684611-1609-4d43-a887-40f21805a2dc\\\",\\\"systemUUID\\\":\\\"3e03dc76-aa4b-4c6f-a4c0-977607dcbe31\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:36Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:36 crc kubenswrapper[4858]: E0320 08:59:36.093915 4858 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 20 08:59:37 crc kubenswrapper[4858]: I0320 08:59:37.069907 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:37 crc kubenswrapper[4858]: I0320 08:59:37.069934 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:37 crc kubenswrapper[4858]: E0320 08:59:37.070106 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:37 crc kubenswrapper[4858]: E0320 08:59:37.070214 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:37 crc kubenswrapper[4858]: I0320 08:59:37.070607 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:37 crc kubenswrapper[4858]: E0320 08:59:37.070814 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:38 crc kubenswrapper[4858]: I0320 08:59:38.069881 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:38 crc kubenswrapper[4858]: E0320 08:59:38.070101 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:39 crc kubenswrapper[4858]: I0320 08:59:39.069711 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:39 crc kubenswrapper[4858]: E0320 08:59:39.069921 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:39 crc kubenswrapper[4858]: I0320 08:59:39.070225 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:39 crc kubenswrapper[4858]: E0320 08:59:39.070365 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:39 crc kubenswrapper[4858]: I0320 08:59:39.070623 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:39 crc kubenswrapper[4858]: E0320 08:59:39.070868 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.069237 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:40 crc kubenswrapper[4858]: E0320 08:59:40.069564 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.096018 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://173ce131ceffab928be8309311b1ec3f92aa8a39fb4411f90cae06b64d9f9f1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.112389 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:41Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e5b603b3a6cb18a4a443415f87663a59a345d97d61afa3c5266ba1c58e0c3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.128306 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae2a40e23046be6ee9a5e68b43cb242de07abbbf5851bf502a6bad33c0673319\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7d6246f1fb32169d4cab14c41e419ef48fc9e82c52caccb5d3d6b7a5b40bb13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.154590 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:27Z\\\",\\\"message\\\":\\\"ces.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0320 08:59:27.867092 7389 services_controller.go:360] Finished syncing service packageserver-service on namespace openshift-operator-lifecycle-manager for network=default : 1.018625ms\\\\nI0320 08:59:27.867561 7389 services_controller.go:452] Built service openshift-kube-controller-manager/kube-controller-manager per-node LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867575 7389 services_controller.go:453] Built service openshift-kube-controller-manager/kube-controller-manager template LB for network=default: []services.LB{}\\\\nI0320 08:59:27.867585 7389 services_controller.go:454] Service openshift-kube-controller-manager/kube-controller-manager for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0320 08:59:27.867590 7389 services_controller.go:356] Processing sync for service openshift-machine-api/machine-api-operator-machine-webhook for network=default\\\\nF0320 08:59:27.867590 7389 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create a\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:59:27Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-htvhn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-dwpzf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.169552 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05c993e7-93a7-468c-a9da-72f4ad40fe80\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02bf33706b70af1d8e3a3015c4df7cf38e7d3044b9deb76ffa00008740e20ebf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6949d65b7416f1781e59e3b04b14188baf7d65af61032256a25c2510206d520e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.187646 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: E0320 08:59:40.190429 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.209184 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca9fdc4c-5d34-40de-bb5d-af6140462f33\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:58:00Z\\\",\\\"message\\\":\\\"le observer\\\\nW0320 08:57:59.689203 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0320 08:57:59.689426 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0320 08:57:59.690174 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3655571493/tls.crt::/tmp/serving-cert-3655571493/tls.key\\\\\\\"\\\\nI0320 08:58:00.032712 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0320 08:58:00.034938 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0320 08:58:00.034957 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0320 08:58:00.034980 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0320 08:58:00.034991 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0320 08:58:00.041745 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0320 08:58:00.041771 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0320 08:58:00.041773 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0320 08:58:00.041776 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0320 08:58:00.041804 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0320 08:58:00.041807 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0320 08:58:00.041810 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0320 08:58:00.041813 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0320 08:58:00.043195 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:59Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.226247 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.242838 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1786fa47-a35b-4b75-bac4-3cfb85a30265\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb8eb490b9f1f082a660f2e1bb0278e64742ed9dc81dc054998da32fae3e8a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee9bfd8db8426bb93610c99b62ab8a044033e3188f808fef581fcb57787618e0\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-20T08:57:25Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0320 08:57:02.150137 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0320 08:57:02.152665 1 observer_polling.go:159] Starting file observer\\\\nI0320 08:57:02.183000 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0320 08:57:02.186998 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0320 08:57:25.758400 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0320 08:57:25.758471 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:57:25Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b12b202361d82a4252119781d9f2f16604d294ba4677dde27fdb29b2df695de5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://94b702c06af9baf43748512ac2ac9ad50c1846422fbddbef0dbb1740cced105d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.257555 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.274308 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-t45zv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e28e3e9c-e621-4e85-af97-c3f48adb269d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb1ce1a031a6f21ec8d56d9b47a99318cf14ee2b1eb7ddb08c376ee2a79602e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb6b940c8d1fc50da0e5dc0bfafc9757afef595fc0b411b939d97985c814030d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11769974d3f586443486b2f2246422ed59cc624a49f12e2a7fbf8b2d20f1002\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://388b3a07bcf95c487f4d11f8208fd75b16e7ce4e66190d39ad480c4d5b04e04d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73935890f1dfb1b177efa3cd3ef0e68f385969f3d3d0d73baa0854a1511208ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b20237f3a08a36e39c93fb24e97f252d9015ff972852683c9adab0962bbd9a9e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c88e014129f395ee6c26a7b00ad03e509be5739c39eae2e3f793b5630d1df48\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:58:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fpfrn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-t45zv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.288589 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-p2cjs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24656c62-314b-4c20-adf1-217d58a95f57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:59:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-20T08:59:24Z\\\",\\\"message\\\":\\\"2026-03-20T08:58:39+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6\\\\n2026-03-20T08:58:39+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c4e4597c-07d3-4141-b9a4-df32be7ed2b6 to /host/opt/cni/bin/\\\\n2026-03-20T08:58:39Z [verbose] multus-daemon started\\\\n2026-03-20T08:58:39Z [verbose] Readiness Indicator file check\\\\n2026-03-20T08:59:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:59:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcxb2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-multus\"/\"multus-p2cjs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.301475 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"11b84afc-479e-4a00-b1d4-936df4acbaa9\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:57:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c0eb4b2b642608370ef84f01a67f4928d8b5d3a484fca26e97bec4477c43ac5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://74743d0091d2ba90a65008b93f91d517b0db1cd677a31c07105b58132013d5e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1900f2e5053080dc7514a18a7fe62c7ec324e8c2b038f031db19d980a8b3f76f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:57:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://119dfe761fcbb47c5bc343581364c032dc7b4b22bc82615b7ec939a0816ff51d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-20T08:57:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-20T08:57:01Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:57:00Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.315411 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-2j2m8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49ca0c4b-5811-4080-b1a6-1c6f02fc5d76\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://10b8dfa63df1c6bb6a14e430fa8393731c0238f773dfb851c83b126032f849db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4qgs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-2j2m8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.330412 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"75db206b-1ea7-4295-85ae-10309c438903\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9743c3c03da94f540e96e1a39838cb2865d39f479757d8dd7216617f9d5d6bdf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4c701d244ab30049230b2f1417e59e685b78af71d60dbc789d1a475c2a99beaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hvzmh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:49Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-22d5l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.343733 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-kvlch" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:51Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rk4cj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:51Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-kvlch\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.356155 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mwh2v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"da8fcd26-7c6c-4a53-a6c6-dadde0238068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f0028f3a7f99265556b18ee0f6832157edf986f3e646229be2d659cc73965eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8q64d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mwh2v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:40 crc kubenswrapper[4858]: I0320 08:59:40.368554 4858 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"584bd2e0-0786-4137-9674-790c8fb680c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-20T08:58:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6ad1fa7050b0e9e8f2dde9b0e01d84bb2528dc8386da3e63f5cca7981b1b8bf4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-20T08:58:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-znr6q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-20T08:58:37Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w6t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-20T08:59:40Z is after 2025-08-24T17:21:41Z" Mar 20 08:59:41 crc kubenswrapper[4858]: I0320 08:59:41.069521 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:41 crc kubenswrapper[4858]: I0320 08:59:41.069537 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:41 crc kubenswrapper[4858]: I0320 08:59:41.069554 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:41 crc kubenswrapper[4858]: E0320 08:59:41.070148 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:41 crc kubenswrapper[4858]: E0320 08:59:41.070228 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:41 crc kubenswrapper[4858]: E0320 08:59:41.070413 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:41 crc kubenswrapper[4858]: I0320 08:59:41.070691 4858 scope.go:117] "RemoveContainer" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" Mar 20 08:59:41 crc kubenswrapper[4858]: E0320 08:59:41.071002 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:59:42 crc kubenswrapper[4858]: I0320 08:59:42.069371 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:42 crc kubenswrapper[4858]: E0320 08:59:42.069669 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:43 crc kubenswrapper[4858]: I0320 08:59:43.069796 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:43 crc kubenswrapper[4858]: E0320 08:59:43.070491 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:43 crc kubenswrapper[4858]: I0320 08:59:43.069859 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:43 crc kubenswrapper[4858]: I0320 08:59:43.069835 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:43 crc kubenswrapper[4858]: E0320 08:59:43.070562 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:43 crc kubenswrapper[4858]: E0320 08:59:43.070869 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:44 crc kubenswrapper[4858]: I0320 08:59:44.069141 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:44 crc kubenswrapper[4858]: E0320 08:59:44.069407 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:45 crc kubenswrapper[4858]: I0320 08:59:45.069129 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:45 crc kubenswrapper[4858]: I0320 08:59:45.069166 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:45 crc kubenswrapper[4858]: E0320 08:59:45.069299 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:45 crc kubenswrapper[4858]: I0320 08:59:45.069395 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:45 crc kubenswrapper[4858]: E0320 08:59:45.069579 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:45 crc kubenswrapper[4858]: E0320 08:59:45.069670 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:45 crc kubenswrapper[4858]: E0320 08:59:45.192051 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.069950 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:46 crc kubenswrapper[4858]: E0320 08:59:46.070251 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.284778 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.284845 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.284862 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.284892 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.284912 4858 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-20T08:59:46Z","lastTransitionTime":"2026-03-20T08:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.346167 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f"] Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.346619 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.348734 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.349163 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.349460 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.350128 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.397226 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=75.397194682 podStartE2EDuration="1m15.397194682s" podCreationTimestamp="2026-03-20 08:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.37112317 +0000 UTC m=+167.691541377" watchObservedRunningTime="2026-03-20 08:59:46.397194682 +0000 UTC m=+167.717612879" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.408932 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.408983 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.409025 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.409058 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.409659 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.429053 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=40.429028171 podStartE2EDuration="40.429028171s" podCreationTimestamp="2026-03-20 08:59:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.416439041 +0000 UTC m=+167.736857238" watchObservedRunningTime="2026-03-20 08:59:46.429028171 +0000 UTC m=+167.749446368" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.445602 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-t45zv" podStartSLOduration=111.445573948 podStartE2EDuration="1m51.445573948s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.444939106 +0000 UTC m=+167.765357303" watchObservedRunningTime="2026-03-20 08:59:46.445573948 +0000 UTC m=+167.765992185" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.459801 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-p2cjs" podStartSLOduration=111.459769713 podStartE2EDuration="1m51.459769713s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.458973756 +0000 UTC m=+167.779392013" watchObservedRunningTime="2026-03-20 08:59:46.459769713 +0000 UTC m=+167.780187910" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.474935 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=42.474908222 podStartE2EDuration="42.474908222s" podCreationTimestamp="2026-03-20 08:59:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.47429753 +0000 UTC m=+167.794715727" watchObservedRunningTime="2026-03-20 08:59:46.474908222 +0000 UTC m=+167.795326459" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.486209 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-2j2m8" podStartSLOduration=111.486178028 podStartE2EDuration="1m51.486178028s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.485520114 +0000 UTC m=+167.805938341" watchObservedRunningTime="2026-03-20 08:59:46.486178028 +0000 UTC m=+167.806596225" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.499651 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-22d5l" podStartSLOduration=111.499629457 podStartE2EDuration="1m51.499629457s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.498965835 +0000 UTC m=+167.819384092" watchObservedRunningTime="2026-03-20 08:59:46.499629457 +0000 UTC m=+167.820047654" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.510558 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.510637 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.510715 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.510753 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.510802 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.510943 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.510933 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.512432 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-service-ca\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.523550 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.538175 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mwh2v" podStartSLOduration=111.538143485 podStartE2EDuration="1m51.538143485s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.536576252 +0000 UTC m=+167.856994459" watchObservedRunningTime="2026-03-20 08:59:46.538143485 +0000 UTC m=+167.858561692" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.540739 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-w5x7f\" (UID: \"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.571401 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podStartSLOduration=111.571371612 podStartE2EDuration="1m51.571371612s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.555359084 +0000 UTC m=+167.875777281" watchObservedRunningTime="2026-03-20 08:59:46.571371612 +0000 UTC m=+167.891789829" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.660408 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" Mar 20 08:59:46 crc kubenswrapper[4858]: I0320 08:59:46.676797 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=75.67677188 podStartE2EDuration="1m15.67677188s" podCreationTimestamp="2026-03-20 08:58:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:46.652873081 +0000 UTC m=+167.973291278" watchObservedRunningTime="2026-03-20 08:59:46.67677188 +0000 UTC m=+167.997190077" Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.069948 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.069964 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.070001 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:47 crc kubenswrapper[4858]: E0320 08:59:47.070307 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:47 crc kubenswrapper[4858]: E0320 08:59:47.070460 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:47 crc kubenswrapper[4858]: E0320 08:59:47.070725 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.097976 4858 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.110256 4858 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.130221 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" event={"ID":"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8","Type":"ContainerStarted","Data":"6d00a8d20b1b8766aea5e7fc8764f022faa0c0d32080f528bfe11a4527e08a7d"} Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.130293 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" event={"ID":"d2cc1f75-2f26-4e26-8ad1-5b5a27dfb5e8","Type":"ContainerStarted","Data":"08a8dfe994e5ebeb2c45b465fd0f539c82360fde70f137aab04605aee95496a2"} Mar 20 08:59:47 crc kubenswrapper[4858]: I0320 08:59:47.149612 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-w5x7f" podStartSLOduration=112.149582498 podStartE2EDuration="1m52.149582498s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:47.149015009 +0000 UTC m=+168.469433266" watchObservedRunningTime="2026-03-20 08:59:47.149582498 +0000 UTC m=+168.470000695" Mar 20 08:59:48 crc kubenswrapper[4858]: I0320 08:59:48.069415 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:48 crc kubenswrapper[4858]: E0320 08:59:48.069832 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:48 crc kubenswrapper[4858]: I0320 08:59:48.095479 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 20 08:59:49 crc kubenswrapper[4858]: I0320 08:59:49.070022 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:49 crc kubenswrapper[4858]: E0320 08:59:49.070601 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:49 crc kubenswrapper[4858]: I0320 08:59:49.070172 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:49 crc kubenswrapper[4858]: E0320 08:59:49.070834 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:49 crc kubenswrapper[4858]: I0320 08:59:49.070145 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:49 crc kubenswrapper[4858]: E0320 08:59:49.071042 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:50 crc kubenswrapper[4858]: I0320 08:59:50.070228 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:50 crc kubenswrapper[4858]: E0320 08:59:50.071886 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:50 crc kubenswrapper[4858]: I0320 08:59:50.107988 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=2.107963074 podStartE2EDuration="2.107963074s" podCreationTimestamp="2026-03-20 08:59:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 08:59:50.107033471 +0000 UTC m=+171.427451718" watchObservedRunningTime="2026-03-20 08:59:50.107963074 +0000 UTC m=+171.428381271" Mar 20 08:59:50 crc kubenswrapper[4858]: E0320 08:59:50.192534 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:51 crc kubenswrapper[4858]: I0320 08:59:51.070022 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:51 crc kubenswrapper[4858]: I0320 08:59:51.070056 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:51 crc kubenswrapper[4858]: E0320 08:59:51.070290 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:51 crc kubenswrapper[4858]: I0320 08:59:51.070388 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:51 crc kubenswrapper[4858]: E0320 08:59:51.070577 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:51 crc kubenswrapper[4858]: E0320 08:59:51.071015 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:52 crc kubenswrapper[4858]: I0320 08:59:52.070263 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:52 crc kubenswrapper[4858]: E0320 08:59:52.070439 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:52 crc kubenswrapper[4858]: I0320 08:59:52.071269 4858 scope.go:117] "RemoveContainer" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" Mar 20 08:59:52 crc kubenswrapper[4858]: E0320 08:59:52.071641 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 08:59:53 crc kubenswrapper[4858]: I0320 08:59:53.069289 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:53 crc kubenswrapper[4858]: I0320 08:59:53.069406 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:53 crc kubenswrapper[4858]: I0320 08:59:53.069416 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:53 crc kubenswrapper[4858]: E0320 08:59:53.069557 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:53 crc kubenswrapper[4858]: E0320 08:59:53.069709 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:53 crc kubenswrapper[4858]: E0320 08:59:53.069918 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:54 crc kubenswrapper[4858]: I0320 08:59:54.069219 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:54 crc kubenswrapper[4858]: E0320 08:59:54.070000 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:55 crc kubenswrapper[4858]: I0320 08:59:55.069735 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:55 crc kubenswrapper[4858]: I0320 08:59:55.069790 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:55 crc kubenswrapper[4858]: I0320 08:59:55.069799 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:55 crc kubenswrapper[4858]: E0320 08:59:55.070575 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:55 crc kubenswrapper[4858]: E0320 08:59:55.070640 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:55 crc kubenswrapper[4858]: E0320 08:59:55.071489 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:55 crc kubenswrapper[4858]: E0320 08:59:55.195037 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 08:59:55 crc kubenswrapper[4858]: I0320 08:59:55.718330 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:55 crc kubenswrapper[4858]: E0320 08:59:55.718545 4858 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:55 crc kubenswrapper[4858]: E0320 08:59:55.718629 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs podName:eb1ef726-a1a8-4efe-bdcc-33fba0e077ea nodeName:}" failed. No retries permitted until 2026-03-20 09:00:59.718607638 +0000 UTC m=+241.039025835 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs") pod "network-metrics-daemon-kvlch" (UID: "eb1ef726-a1a8-4efe-bdcc-33fba0e077ea") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 20 08:59:56 crc kubenswrapper[4858]: I0320 08:59:56.069229 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:56 crc kubenswrapper[4858]: E0320 08:59:56.069499 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:57 crc kubenswrapper[4858]: I0320 08:59:57.069849 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:57 crc kubenswrapper[4858]: I0320 08:59:57.069925 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:57 crc kubenswrapper[4858]: E0320 08:59:57.069998 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:57 crc kubenswrapper[4858]: E0320 08:59:57.070099 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 08:59:57 crc kubenswrapper[4858]: I0320 08:59:57.069865 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:57 crc kubenswrapper[4858]: E0320 08:59:57.070179 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:58 crc kubenswrapper[4858]: I0320 08:59:58.070074 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 08:59:58 crc kubenswrapper[4858]: E0320 08:59:58.070296 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 08:59:59 crc kubenswrapper[4858]: I0320 08:59:59.069295 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 08:59:59 crc kubenswrapper[4858]: I0320 08:59:59.069295 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 08:59:59 crc kubenswrapper[4858]: I0320 08:59:59.069514 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 08:59:59 crc kubenswrapper[4858]: E0320 08:59:59.069652 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 08:59:59 crc kubenswrapper[4858]: E0320 08:59:59.069805 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 08:59:59 crc kubenswrapper[4858]: E0320 08:59:59.069936 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:00 crc kubenswrapper[4858]: I0320 09:00:00.070712 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:00 crc kubenswrapper[4858]: E0320 09:00:00.072243 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:00 crc kubenswrapper[4858]: E0320 09:00:00.195692 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:01 crc kubenswrapper[4858]: I0320 09:00:01.069723 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:01 crc kubenswrapper[4858]: I0320 09:00:01.069940 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:01 crc kubenswrapper[4858]: I0320 09:00:01.070276 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:01 crc kubenswrapper[4858]: E0320 09:00:01.070907 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:01 crc kubenswrapper[4858]: E0320 09:00:01.071065 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:01 crc kubenswrapper[4858]: E0320 09:00:01.071151 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:02 crc kubenswrapper[4858]: I0320 09:00:02.069921 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:02 crc kubenswrapper[4858]: E0320 09:00:02.071039 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:03 crc kubenswrapper[4858]: I0320 09:00:03.069860 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:03 crc kubenswrapper[4858]: I0320 09:00:03.069931 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:03 crc kubenswrapper[4858]: I0320 09:00:03.070001 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:03 crc kubenswrapper[4858]: E0320 09:00:03.070042 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:03 crc kubenswrapper[4858]: E0320 09:00:03.070130 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:03 crc kubenswrapper[4858]: E0320 09:00:03.070233 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:04 crc kubenswrapper[4858]: I0320 09:00:04.069803 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:04 crc kubenswrapper[4858]: E0320 09:00:04.070014 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:05 crc kubenswrapper[4858]: I0320 09:00:05.069748 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:05 crc kubenswrapper[4858]: I0320 09:00:05.069803 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:05 crc kubenswrapper[4858]: E0320 09:00:05.069938 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:05 crc kubenswrapper[4858]: I0320 09:00:05.070059 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:05 crc kubenswrapper[4858]: E0320 09:00:05.070171 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:05 crc kubenswrapper[4858]: E0320 09:00:05.070280 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:05 crc kubenswrapper[4858]: I0320 09:00:05.071695 4858 scope.go:117] "RemoveContainer" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" Mar 20 09:00:05 crc kubenswrapper[4858]: E0320 09:00:05.072021 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-dwpzf_openshift-ovn-kubernetes(21fd7c33-ddc7-4a05-a922-472eb8ccd4e1)\"" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" Mar 20 09:00:05 crc kubenswrapper[4858]: E0320 09:00:05.196851 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:06 crc kubenswrapper[4858]: I0320 09:00:06.069736 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:06 crc kubenswrapper[4858]: E0320 09:00:06.069917 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:07 crc kubenswrapper[4858]: I0320 09:00:07.069880 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:07 crc kubenswrapper[4858]: I0320 09:00:07.069959 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:07 crc kubenswrapper[4858]: I0320 09:00:07.070005 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:07 crc kubenswrapper[4858]: E0320 09:00:07.070130 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:07 crc kubenswrapper[4858]: E0320 09:00:07.070356 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:07 crc kubenswrapper[4858]: E0320 09:00:07.070471 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:08 crc kubenswrapper[4858]: I0320 09:00:08.069433 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:08 crc kubenswrapper[4858]: E0320 09:00:08.069672 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:09 crc kubenswrapper[4858]: I0320 09:00:09.069135 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:09 crc kubenswrapper[4858]: I0320 09:00:09.069214 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:09 crc kubenswrapper[4858]: I0320 09:00:09.069230 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:09 crc kubenswrapper[4858]: E0320 09:00:09.069396 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:09 crc kubenswrapper[4858]: E0320 09:00:09.069532 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:09 crc kubenswrapper[4858]: E0320 09:00:09.069697 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:10 crc kubenswrapper[4858]: I0320 09:00:10.070026 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:10 crc kubenswrapper[4858]: E0320 09:00:10.075715 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:10 crc kubenswrapper[4858]: E0320 09:00:10.197835 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:10 crc kubenswrapper[4858]: I0320 09:00:10.210025 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/1.log" Mar 20 09:00:10 crc kubenswrapper[4858]: I0320 09:00:10.210693 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/0.log" Mar 20 09:00:10 crc kubenswrapper[4858]: I0320 09:00:10.210864 4858 generic.go:334] "Generic (PLEG): container finished" podID="24656c62-314b-4c20-adf1-217d58a95f57" containerID="3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020" exitCode=1 Mar 20 09:00:10 crc kubenswrapper[4858]: I0320 09:00:10.210946 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerDied","Data":"3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020"} Mar 20 09:00:10 crc kubenswrapper[4858]: I0320 09:00:10.211153 4858 scope.go:117] "RemoveContainer" containerID="574a9d8fb05f39a789a3ff977ba784a0ece19d2172f614acc86b9ef5cb462ce0" Mar 20 09:00:10 crc kubenswrapper[4858]: I0320 09:00:10.211916 4858 scope.go:117] "RemoveContainer" containerID="3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020" Mar 20 09:00:10 crc kubenswrapper[4858]: E0320 09:00:10.212278 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-p2cjs_openshift-multus(24656c62-314b-4c20-adf1-217d58a95f57)\"" pod="openshift-multus/multus-p2cjs" podUID="24656c62-314b-4c20-adf1-217d58a95f57" Mar 20 09:00:11 crc kubenswrapper[4858]: I0320 09:00:11.069711 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:11 crc kubenswrapper[4858]: I0320 09:00:11.069800 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:11 crc kubenswrapper[4858]: I0320 09:00:11.069754 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:11 crc kubenswrapper[4858]: E0320 09:00:11.070007 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:11 crc kubenswrapper[4858]: E0320 09:00:11.070127 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:11 crc kubenswrapper[4858]: E0320 09:00:11.070548 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:11 crc kubenswrapper[4858]: I0320 09:00:11.216913 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/1.log" Mar 20 09:00:12 crc kubenswrapper[4858]: I0320 09:00:12.069666 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:12 crc kubenswrapper[4858]: E0320 09:00:12.069984 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:13 crc kubenswrapper[4858]: I0320 09:00:13.069171 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:13 crc kubenswrapper[4858]: I0320 09:00:13.069270 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:13 crc kubenswrapper[4858]: I0320 09:00:13.069203 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:13 crc kubenswrapper[4858]: E0320 09:00:13.069517 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:13 crc kubenswrapper[4858]: E0320 09:00:13.069710 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:13 crc kubenswrapper[4858]: E0320 09:00:13.069915 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:14 crc kubenswrapper[4858]: I0320 09:00:14.069471 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:14 crc kubenswrapper[4858]: E0320 09:00:14.069682 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:15 crc kubenswrapper[4858]: I0320 09:00:15.069301 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:15 crc kubenswrapper[4858]: I0320 09:00:15.069302 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:15 crc kubenswrapper[4858]: E0320 09:00:15.069512 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:15 crc kubenswrapper[4858]: E0320 09:00:15.069644 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:15 crc kubenswrapper[4858]: I0320 09:00:15.070006 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:15 crc kubenswrapper[4858]: E0320 09:00:15.070109 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:15 crc kubenswrapper[4858]: E0320 09:00:15.199528 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:16 crc kubenswrapper[4858]: I0320 09:00:16.069985 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:16 crc kubenswrapper[4858]: E0320 09:00:16.070195 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.069709 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.069703 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.070633 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:17 crc kubenswrapper[4858]: E0320 09:00:17.071225 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:17 crc kubenswrapper[4858]: E0320 09:00:17.071376 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:17 crc kubenswrapper[4858]: E0320 09:00:17.071481 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.071873 4858 scope.go:117] "RemoveContainer" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.245502 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/3.log" Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.249255 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerStarted","Data":"96b99b2166595d06311919f39bcf5f3bcd3dd03439156ccbcfd3b92bcdf473f0"} Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.249934 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 09:00:17 crc kubenswrapper[4858]: I0320 09:00:17.295458 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podStartSLOduration=142.295422318 podStartE2EDuration="2m22.295422318s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:17.293174407 +0000 UTC m=+198.613592644" watchObservedRunningTime="2026-03-20 09:00:17.295422318 +0000 UTC m=+198.615840545" Mar 20 09:00:18 crc kubenswrapper[4858]: I0320 09:00:18.069498 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:18 crc kubenswrapper[4858]: E0320 09:00:18.069793 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:18 crc kubenswrapper[4858]: I0320 09:00:18.088335 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kvlch"] Mar 20 09:00:18 crc kubenswrapper[4858]: I0320 09:00:18.254417 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:18 crc kubenswrapper[4858]: E0320 09:00:18.254510 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:19 crc kubenswrapper[4858]: I0320 09:00:19.070635 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:19 crc kubenswrapper[4858]: I0320 09:00:19.070676 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:19 crc kubenswrapper[4858]: I0320 09:00:19.070690 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:19 crc kubenswrapper[4858]: E0320 09:00:19.071139 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:19 crc kubenswrapper[4858]: E0320 09:00:19.072125 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:19 crc kubenswrapper[4858]: E0320 09:00:19.072620 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:20 crc kubenswrapper[4858]: I0320 09:00:20.069433 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:20 crc kubenswrapper[4858]: E0320 09:00:20.070660 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:20 crc kubenswrapper[4858]: E0320 09:00:20.200917 4858 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:00:21 crc kubenswrapper[4858]: I0320 09:00:21.069058 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:21 crc kubenswrapper[4858]: E0320 09:00:21.069201 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:21 crc kubenswrapper[4858]: I0320 09:00:21.069280 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:21 crc kubenswrapper[4858]: I0320 09:00:21.069455 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:21 crc kubenswrapper[4858]: E0320 09:00:21.069492 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:21 crc kubenswrapper[4858]: E0320 09:00:21.069651 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:22 crc kubenswrapper[4858]: I0320 09:00:22.069899 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:22 crc kubenswrapper[4858]: E0320 09:00:22.070300 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:22 crc kubenswrapper[4858]: I0320 09:00:22.070588 4858 scope.go:117] "RemoveContainer" containerID="3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020" Mar 20 09:00:22 crc kubenswrapper[4858]: I0320 09:00:22.270792 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/1.log" Mar 20 09:00:22 crc kubenswrapper[4858]: I0320 09:00:22.270872 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerStarted","Data":"ba981e9a0ce9b14170b3eabfd0ccf4d14c784ab266253c7e4f5a08575832c5ac"} Mar 20 09:00:23 crc kubenswrapper[4858]: I0320 09:00:23.069650 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:23 crc kubenswrapper[4858]: E0320 09:00:23.069769 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:23 crc kubenswrapper[4858]: I0320 09:00:23.069675 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:23 crc kubenswrapper[4858]: E0320 09:00:23.069833 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:23 crc kubenswrapper[4858]: I0320 09:00:23.069661 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:23 crc kubenswrapper[4858]: E0320 09:00:23.070379 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:24 crc kubenswrapper[4858]: I0320 09:00:24.069863 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:24 crc kubenswrapper[4858]: E0320 09:00:24.070464 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-kvlch" podUID="eb1ef726-a1a8-4efe-bdcc-33fba0e077ea" Mar 20 09:00:25 crc kubenswrapper[4858]: I0320 09:00:25.069457 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:25 crc kubenswrapper[4858]: I0320 09:00:25.069579 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:25 crc kubenswrapper[4858]: E0320 09:00:25.069961 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 20 09:00:25 crc kubenswrapper[4858]: I0320 09:00:25.069623 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:25 crc kubenswrapper[4858]: E0320 09:00:25.070067 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 20 09:00:25 crc kubenswrapper[4858]: E0320 09:00:25.070287 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.069434 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.076710 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.076989 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.901624 4858 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.952282 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c5kzc"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.953372 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.953512 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hsp4b"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.954173 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.968094 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5grwr"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.969210 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.969928 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.970725 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.975800 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.976230 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.976397 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.986105 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.988636 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.988873 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.989001 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.989135 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.989296 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.990204 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.990379 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.990665 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.990719 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.990904 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.991160 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.991187 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.991259 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.991347 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.991704 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.991884 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992089 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992227 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992299 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992361 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992425 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992475 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992568 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992437 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992835 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.993426 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992682 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992762 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992771 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992787 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992829 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.992869 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.993409 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.993474 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.997171 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.997498 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cwnt6"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.998045 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2c66r"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.998506 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vw55r"] Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.999137 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.999716 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:26 crc kubenswrapper[4858]: I0320 09:00:26.999933 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.002780 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-audit-dir\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.002824 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj6td\" (UniqueName: \"kubernetes.io/projected/039cac36-f4ed-4282-aa07-ee40ad00df93-kube-api-access-rj6td\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.002863 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.002894 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.002929 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-image-import-ca\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.002958 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-encryption-config\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.002979 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-node-pullsecrets\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003004 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-etcd-client\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003038 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039cac36-f4ed-4282-aa07-ee40ad00df93-serving-cert\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003070 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-audit\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003096 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-serving-cert\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003133 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfgvt\" (UniqueName: \"kubernetes.io/projected/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-kube-api-access-xfgvt\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003166 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-client-ca\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003189 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-config\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003207 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-config\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.003233 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-etcd-serving-ca\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.004817 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.005015 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jsp8n"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.005805 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-vtwn4"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.006173 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.006221 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.006702 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.011783 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.014265 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.016372 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-j6mmm"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.016818 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-j6mmm" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.017167 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-wr84h"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.023428 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hsp4b"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.028481 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.028479 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.029796 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.029830 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.030015 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.035009 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.035362 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.045094 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.045187 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.045702 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.045848 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.046053 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.046074 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.046528 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.046536 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.046878 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.046916 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.047084 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.047241 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.047268 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.047634 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.055435 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.055816 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.056156 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.056816 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.056996 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.057238 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.064368 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.064885 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.066451 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-468wl"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067304 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067536 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067567 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067721 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8crqq"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067651 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067795 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067847 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067885 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067739 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.067968 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068027 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068128 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068166 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068202 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068452 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068487 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068538 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068563 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068668 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068789 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068856 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068888 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068924 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068818 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068994 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.068843 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069075 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069092 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069158 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069184 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069518 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069667 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069679 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069693 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069826 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069880 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069898 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.069946 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.070066 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.070222 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.070276 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.071266 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.073784 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.074161 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.074707 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.075080 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.075672 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.077253 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.077887 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.077997 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.078609 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.083573 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.084349 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.084456 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.084795 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.085586 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.086426 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.096352 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.097046 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.087309 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.095564 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.097762 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.097863 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.098183 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.098740 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.098932 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.099108 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.101758 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.104943 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/990e152e-f7bc-4811-bc9a-6954a09b166a-serving-cert\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105009 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b3bd918-04b9-4371-933a-609e9add5512-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105042 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-encryption-config\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105067 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-dir\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105085 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105109 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-service-ca-bundle\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105129 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-node-pullsecrets\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105146 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-etcd-client\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105164 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-default-certificate\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105190 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105208 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/effb9468-b572-4eb1-84df-15e7b0201dbf-serving-cert\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105229 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-config\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105251 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8tvn\" (UniqueName: \"kubernetes.io/projected/c245d181-9680-448c-a0c6-32f5d54811f7-kube-api-access-m8tvn\") pod \"cluster-samples-operator-665b6dd947-xm7ws\" (UID: \"c245d181-9680-448c-a0c6-32f5d54811f7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105276 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-serving-cert\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105296 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039cac36-f4ed-4282-aa07-ee40ad00df93-serving-cert\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105333 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c245d181-9680-448c-a0c6-32f5d54811f7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xm7ws\" (UID: \"c245d181-9680-448c-a0c6-32f5d54811f7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105353 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-trusted-ca-bundle\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105369 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxr8p\" (UniqueName: \"kubernetes.io/projected/990e152e-f7bc-4811-bc9a-6954a09b166a-kube-api-access-gxr8p\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105390 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-config\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105415 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-client\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105434 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-audit\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105451 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-config\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105465 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w92k6\" (UniqueName: \"kubernetes.io/projected/99346cc6-9090-4e06-beb0-d64a92bd2813-kube-api-access-w92k6\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105481 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/830fcf94-999e-4859-a62e-f317fc53eaf6-service-ca-bundle\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105508 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtww9\" (UniqueName: \"kubernetes.io/projected/a9f367b2-d0b3-4a80-933f-68bf11e63791-kube-api-access-vtww9\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105530 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/31262c04-c5c9-4b06-afd8-f005d271819a-machine-approver-tls\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105550 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghvr5\" (UniqueName: \"kubernetes.io/projected/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-kube-api-access-ghvr5\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105569 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59f9f7fc-da90-46ec-b360-1eee512a4416-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105590 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-serving-cert\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105607 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b3bd918-04b9-4371-933a-609e9add5512-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105624 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sg8h\" (UniqueName: \"kubernetes.io/projected/e3892972-3419-4c52-bb2f-993e4a19d813-kube-api-access-6sg8h\") pod \"dns-operator-744455d44c-jsp8n\" (UID: \"e3892972-3419-4c52-bb2f-993e4a19d813\") " pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105642 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-etcd-client\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105657 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105677 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105693 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105717 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105735 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-stats-auth\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105753 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31262c04-c5c9-4b06-afd8-f005d271819a-auth-proxy-config\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105775 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4mgr\" (UniqueName: \"kubernetes.io/projected/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-kube-api-access-m4mgr\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105798 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105818 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-config\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105842 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105862 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105884 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcn9q\" (UniqueName: \"kubernetes.io/projected/e07edf68-41a8-4175-adc0-163e46620ab4-kube-api-access-qcn9q\") pod \"downloads-7954f5f757-j6mmm\" (UID: \"e07edf68-41a8-4175-adc0-163e46620ab4\") " pod="openshift-console/downloads-7954f5f757-j6mmm" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105905 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105926 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shhz7\" (UniqueName: \"kubernetes.io/projected/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-kube-api-access-shhz7\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105946 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-client-ca\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105967 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-images\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.105990 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31262c04-c5c9-4b06-afd8-f005d271819a-config\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106009 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-oauth-config\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106029 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106048 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-trusted-ca\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106076 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxk6j\" (UniqueName: \"kubernetes.io/projected/055f16c2-9ca1-4078-82b9-48aa9a4399ad-kube-api-access-lxk6j\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106097 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfgvt\" (UniqueName: \"kubernetes.io/projected/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-kube-api-access-xfgvt\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106128 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-policies\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106144 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-service-ca\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106163 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v727w\" (UniqueName: \"kubernetes.io/projected/3b3bd918-04b9-4371-933a-609e9add5512-kube-api-access-v727w\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106179 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59f9f7fc-da90-46ec-b360-1eee512a4416-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106206 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-client-ca\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106227 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106249 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-config\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106267 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-config\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106285 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-etcd-serving-ca\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106305 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-metrics-certs\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106345 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-serving-cert\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106371 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/990e152e-f7bc-4811-bc9a-6954a09b166a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106391 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-console-config\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106431 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106448 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3892972-3419-4c52-bb2f-993e4a19d813-metrics-tls\") pod \"dns-operator-744455d44c-jsp8n\" (UID: \"e3892972-3419-4c52-bb2f-993e4a19d813\") " pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106465 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-oauth-serving-cert\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106485 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-config\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106503 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vl4n\" (UniqueName: \"kubernetes.io/projected/830fcf94-999e-4859-a62e-f317fc53eaf6-kube-api-access-9vl4n\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106520 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-service-ca\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106546 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106564 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-serving-cert\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106583 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj6td\" (UniqueName: \"kubernetes.io/projected/039cac36-f4ed-4282-aa07-ee40ad00df93-kube-api-access-rj6td\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106599 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85htn\" (UniqueName: \"kubernetes.io/projected/31262c04-c5c9-4b06-afd8-f005d271819a-kube-api-access-85htn\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.106619 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct5l6\" (UniqueName: \"kubernetes.io/projected/59f9f7fc-da90-46ec-b360-1eee512a4416-kube-api-access-ct5l6\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.124440 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.124576 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-ca\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.124651 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126137 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126208 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-audit-dir\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126270 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-encryption-config\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126374 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-audit-dir\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126419 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126448 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99346cc6-9090-4e06-beb0-d64a92bd2813-serving-cert\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126507 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tszd\" (UniqueName: \"kubernetes.io/projected/effb9468-b572-4eb1-84df-15e7b0201dbf-kube-api-access-7tszd\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126545 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-image-import-ca\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126576 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdvl5\" (UniqueName: \"kubernetes.io/projected/91be84d3-8196-44bb-8a88-e9e6548377a1-kube-api-access-qdvl5\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126605 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-serving-cert\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126631 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/055f16c2-9ca1-4078-82b9-48aa9a4399ad-audit-dir\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.126657 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-audit-policies\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.127899 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.129583 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-image-import-ca\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.131400 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.131944 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-audit\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.132259 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-node-pullsecrets\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.133299 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.133372 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-etcd-serving-ca\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.134072 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-client-ca\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.135199 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.136049 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.136610 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.137837 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-config\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.138558 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-config\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.138994 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-serving-cert\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.141999 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.144045 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039cac36-f4ed-4282-aa07-ee40ad00df93-serving-cert\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.144256 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-etcd-client\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.144756 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.145034 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.145669 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.147120 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.148463 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9rv48"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.149826 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.154879 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.160851 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.161804 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4s5wf"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.161985 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.163357 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.164020 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.164211 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.164333 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.164932 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.165497 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.165926 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-encryption-config\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.165951 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.167819 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566620-gz4tx"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.168366 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.169461 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h64jx"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.170444 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.170938 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c5kzc"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.171987 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.176047 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.176429 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.177095 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5grwr"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.188842 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.190425 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jsp8n"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.191463 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.192548 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.193788 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cwnt6"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.194907 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-j6mmm"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.195758 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.195985 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.196999 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8crqq"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.197793 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2c66r"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.200947 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wr84h"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.201536 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.206764 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.207959 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.208913 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-b58fr"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.210722 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.211076 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.212616 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.213761 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-468wl"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.215668 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.216330 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.217468 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.218584 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.219778 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9rv48"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.220855 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.222999 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.223810 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.224928 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vw55r"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.226043 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.227216 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59f9f7fc-da90-46ec-b360-1eee512a4416-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.227330 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghvr5\" (UniqueName: \"kubernetes.io/projected/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-kube-api-access-ghvr5\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.227427 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b3bd918-04b9-4371-933a-609e9add5512-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.227503 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sg8h\" (UniqueName: \"kubernetes.io/projected/e3892972-3419-4c52-bb2f-993e4a19d813-kube-api-access-6sg8h\") pod \"dns-operator-744455d44c-jsp8n\" (UID: \"e3892972-3419-4c52-bb2f-993e4a19d813\") " pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.227582 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.227651 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.227734 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.228523 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.228618 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.228695 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-stats-auth\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.228818 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31262c04-c5c9-4b06-afd8-f005d271819a-auth-proxy-config\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.228933 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-etcd-client\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.229003 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.229078 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-config\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.229158 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.229231 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4mgr\" (UniqueName: \"kubernetes.io/projected/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-kube-api-access-m4mgr\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.229308 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.230955 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcn9q\" (UniqueName: \"kubernetes.io/projected/e07edf68-41a8-4175-adc0-163e46620ab4-kube-api-access-qcn9q\") pod \"downloads-7954f5f757-j6mmm\" (UID: \"e07edf68-41a8-4175-adc0-163e46620ab4\") " pod="openshift-console/downloads-7954f5f757-j6mmm" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231060 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.228755 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231225 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shhz7\" (UniqueName: \"kubernetes.io/projected/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-kube-api-access-shhz7\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231298 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-proxy-tls\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231720 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-images\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231759 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31262c04-c5c9-4b06-afd8-f005d271819a-config\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231766 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-config\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231782 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-client-ca\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231835 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231864 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.230238 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231901 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-trusted-ca\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.230375 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31262c04-c5c9-4b06-afd8-f005d271819a-auth-proxy-config\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.230516 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231218 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-gz4tx"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232090 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h64jx"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.231069 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232107 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rzvjq"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232362 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxk6j\" (UniqueName: \"kubernetes.io/projected/055f16c2-9ca1-4078-82b9-48aa9a4399ad-kube-api-access-lxk6j\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232672 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-oauth-config\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232712 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-service-ca\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232757 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v727w\" (UniqueName: \"kubernetes.io/projected/3b3bd918-04b9-4371-933a-609e9add5512-kube-api-access-v727w\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232779 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59f9f7fc-da90-46ec-b360-1eee512a4416-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232804 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-policies\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232826 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232848 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/960976f4-1bea-423a-b4fc-09b08a60ba0d-serving-cert\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232888 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-metrics-certs\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232911 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-serving-cert\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232934 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwfzf\" (UniqueName: \"kubernetes.io/projected/960976f4-1bea-423a-b4fc-09b08a60ba0d-kube-api-access-dwfzf\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232954 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-console-config\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232979 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233000 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3892972-3419-4c52-bb2f-993e4a19d813-metrics-tls\") pod \"dns-operator-744455d44c-jsp8n\" (UID: \"e3892972-3419-4c52-bb2f-993e4a19d813\") " pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233019 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/990e152e-f7bc-4811-bc9a-6954a09b166a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233037 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-config\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233044 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-client-ca\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233059 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-oauth-serving-cert\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233100 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vl4n\" (UniqueName: \"kubernetes.io/projected/830fcf94-999e-4859-a62e-f317fc53eaf6-kube-api-access-9vl4n\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233123 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-service-ca\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233157 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233182 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-images\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233206 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-serving-cert\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233229 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85htn\" (UniqueName: \"kubernetes.io/projected/31262c04-c5c9-4b06-afd8-f005d271819a-kube-api-access-85htn\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233257 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct5l6\" (UniqueName: \"kubernetes.io/projected/59f9f7fc-da90-46ec-b360-1eee512a4416-kube-api-access-ct5l6\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233278 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233272 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-images\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233325 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-ca\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236152 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-encryption-config\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236187 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99346cc6-9090-4e06-beb0-d64a92bd2813-serving-cert\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236242 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swq66\" (UniqueName: \"kubernetes.io/projected/69e5a3c1-6c66-42f5-a122-63c4d2838aca-kube-api-access-swq66\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233483 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236304 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6scs\" (UniqueName: \"kubernetes.io/projected/74fe10ec-a162-4c93-b2d3-1a80745e7fcc-kube-api-access-v6scs\") pod \"auto-csr-approver-29566620-gz4tx\" (UID: \"74fe10ec-a162-4c93-b2d3-1a80745e7fcc\") " pod="openshift-infra/auto-csr-approver-29566620-gz4tx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236362 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tszd\" (UniqueName: \"kubernetes.io/projected/effb9468-b572-4eb1-84df-15e7b0201dbf-kube-api-access-7tszd\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236391 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdvl5\" (UniqueName: \"kubernetes.io/projected/91be84d3-8196-44bb-8a88-e9e6548377a1-kube-api-access-qdvl5\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236432 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/055f16c2-9ca1-4078-82b9-48aa9a4399ad-audit-dir\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236465 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-audit-policies\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236495 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-serving-cert\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236527 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b3bd918-04b9-4371-933a-609e9add5512-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236553 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/990e152e-f7bc-4811-bc9a-6954a09b166a-serving-cert\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233378 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.235568 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-etcd-client\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236619 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tbsv5"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236640 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-dir\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.233717 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.234047 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-trusted-ca\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.235111 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-oauth-serving-cert\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.232946 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31262c04-c5c9-4b06-afd8-f005d271819a-config\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236847 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/055f16c2-9ca1-4078-82b9-48aa9a4399ad-audit-dir\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.235379 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-policies\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.237634 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.237877 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b3bd918-04b9-4371-933a-609e9add5512-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.237904 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.238012 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/990e152e-f7bc-4811-bc9a-6954a09b166a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.238335 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.238339 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59f9f7fc-da90-46ec-b360-1eee512a4416-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.238500 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236585 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-dir\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239019 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239114 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/69e5a3c1-6c66-42f5-a122-63c4d2838aca-profile-collector-cert\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239170 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.237950 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/055f16c2-9ca1-4078-82b9-48aa9a4399ad-audit-policies\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236054 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-console-config\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239385 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-service-ca-bundle\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239511 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-default-certificate\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239641 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q65g9\" (UniqueName: \"kubernetes.io/projected/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-kube-api-access-q65g9\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239699 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b3bd918-04b9-4371-933a-609e9add5512-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.234815 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59f9f7fc-da90-46ec-b360-1eee512a4416-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239513 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.239946 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.240037 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/effb9468-b572-4eb1-84df-15e7b0201dbf-serving-cert\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.240141 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-service-ca-bundle\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.235783 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.236044 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-config\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241457 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-config\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241623 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960976f4-1bea-423a-b4fc-09b08a60ba0d-config\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241712 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-serving-cert\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241789 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/69e5a3c1-6c66-42f5-a122-63c4d2838aca-srv-cert\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241875 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8tvn\" (UniqueName: \"kubernetes.io/projected/c245d181-9680-448c-a0c6-32f5d54811f7-kube-api-access-m8tvn\") pod \"cluster-samples-operator-665b6dd947-xm7ws\" (UID: \"c245d181-9680-448c-a0c6-32f5d54811f7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241800 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241963 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-trusted-ca-bundle\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242294 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxr8p\" (UniqueName: \"kubernetes.io/projected/990e152e-f7bc-4811-bc9a-6954a09b166a-kube-api-access-gxr8p\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242101 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-config\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242386 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-config\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242556 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-client\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242598 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c245d181-9680-448c-a0c6-32f5d54811f7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xm7ws\" (UID: \"c245d181-9680-448c-a0c6-32f5d54811f7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242634 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w92k6\" (UniqueName: \"kubernetes.io/projected/99346cc6-9090-4e06-beb0-d64a92bd2813-kube-api-access-w92k6\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242694 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/830fcf94-999e-4859-a62e-f317fc53eaf6-service-ca-bundle\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242725 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtww9\" (UniqueName: \"kubernetes.io/projected/a9f367b2-d0b3-4a80-933f-68bf11e63791-kube-api-access-vtww9\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242755 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-config\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.242790 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/31262c04-c5c9-4b06-afd8-f005d271819a-machine-approver-tls\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.241911 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.243729 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e3892972-3419-4c52-bb2f-993e4a19d813-metrics-tls\") pod \"dns-operator-744455d44c-jsp8n\" (UID: \"e3892972-3419-4c52-bb2f-993e4a19d813\") " pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.244132 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/990e152e-f7bc-4811-bc9a-6954a09b166a-serving-cert\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.244511 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-trusted-ca-bundle\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.244683 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/830fcf94-999e-4859-a62e-f317fc53eaf6-service-ca-bundle\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.244864 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-config\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.244997 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-service-ca\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.245222 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-encryption-config\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.245260 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-metrics-certs\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.245751 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/055f16c2-9ca1-4078-82b9-48aa9a4399ad-serving-cert\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.245920 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-serving-cert\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.246192 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/31262c04-c5c9-4b06-afd8-f005d271819a-machine-approver-tls\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.246918 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.247095 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.247788 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/c245d181-9680-448c-a0c6-32f5d54811f7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xm7ws\" (UID: \"c245d181-9680-448c-a0c6-32f5d54811f7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.250866 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.250927 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-stats-auth\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.250950 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b58fr"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.251379 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-serving-cert\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.251993 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/830fcf94-999e-4859-a62e-f317fc53eaf6-default-certificate\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.253213 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.253567 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/effb9468-b572-4eb1-84df-15e7b0201dbf-serving-cert\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.253745 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-serving-cert\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.253903 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-oauth-config\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.256786 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.259813 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.261136 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4s5wf"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.262356 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tbsv5"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.263510 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-c47vz"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.264669 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.264820 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c47vz"] Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.278357 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.307608 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.309501 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-config\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.330260 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.334134 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-ca\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.336332 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344282 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6scs\" (UniqueName: \"kubernetes.io/projected/74fe10ec-a162-4c93-b2d3-1a80745e7fcc-kube-api-access-v6scs\") pod \"auto-csr-approver-29566620-gz4tx\" (UID: \"74fe10ec-a162-4c93-b2d3-1a80745e7fcc\") " pod="openshift-infra/auto-csr-approver-29566620-gz4tx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344332 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swq66\" (UniqueName: \"kubernetes.io/projected/69e5a3c1-6c66-42f5-a122-63c4d2838aca-kube-api-access-swq66\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344375 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/69e5a3c1-6c66-42f5-a122-63c4d2838aca-profile-collector-cert\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344406 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960976f4-1bea-423a-b4fc-09b08a60ba0d-config\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344429 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q65g9\" (UniqueName: \"kubernetes.io/projected/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-kube-api-access-q65g9\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344456 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/69e5a3c1-6c66-42f5-a122-63c4d2838aca-srv-cert\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344524 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344574 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-proxy-tls\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344597 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-service-ca\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344638 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/960976f4-1bea-423a-b4fc-09b08a60ba0d-serving-cert\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344670 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwfzf\" (UniqueName: \"kubernetes.io/projected/960976f4-1bea-423a-b4fc-09b08a60ba0d-kube-api-access-dwfzf\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.344708 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-images\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.345424 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.356670 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.359398 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99346cc6-9090-4e06-beb0-d64a92bd2813-serving-cert\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.375780 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.386880 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99346cc6-9090-4e06-beb0-d64a92bd2813-etcd-client\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.396395 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.417022 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.436187 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.462117 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.475978 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.496701 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.516746 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.536185 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.556369 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.577041 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.596356 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.616449 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.637214 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.658247 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.671178 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/69e5a3c1-6c66-42f5-a122-63c4d2838aca-srv-cert\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.676986 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.696520 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.725920 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.738628 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.750135 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/69e5a3c1-6c66-42f5-a122-63c4d2838aca-profile-collector-cert\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.756484 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.776044 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.798016 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.818181 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.837204 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.858357 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.876646 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.898183 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.917913 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.937268 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.946057 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-images\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.957905 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.977043 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 09:00:27 crc kubenswrapper[4858]: I0320 09:00:27.996474 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.010468 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-proxy-tls\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.017265 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.076972 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.077392 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfgvt\" (UniqueName: \"kubernetes.io/projected/84a3f478-6ec6-4ef6-983e-acdeeb3b475c-kube-api-access-xfgvt\") pod \"apiserver-76f77b778f-c5kzc\" (UID: \"84a3f478-6ec6-4ef6-983e-acdeeb3b475c\") " pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.096434 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.133978 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj6td\" (UniqueName: \"kubernetes.io/projected/039cac36-f4ed-4282-aa07-ee40ad00df93-kube-api-access-rj6td\") pod \"controller-manager-879f6c89f-hsp4b\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.136661 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.155053 4858 request.go:700] Waited for 1.018112571s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dcollect-profiles-dockercfg-kzf4t&limit=500&resourceVersion=0 Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.156707 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.177199 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.192791 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.200769 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.216532 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.236838 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.237128 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.259477 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.277369 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.296391 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.319003 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.337555 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: E0320 09:00:28.345009 4858 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Mar 20 09:00:28 crc kubenswrapper[4858]: E0320 09:00:28.345178 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/960976f4-1bea-423a-b4fc-09b08a60ba0d-serving-cert podName:960976f4-1bea-423a-b4fc-09b08a60ba0d nodeName:}" failed. No retries permitted until 2026-03-20 09:00:28.84512784 +0000 UTC m=+210.165546087 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/960976f4-1bea-423a-b4fc-09b08a60ba0d-serving-cert") pod "service-ca-operator-777779d784-fzvwc" (UID: "960976f4-1bea-423a-b4fc-09b08a60ba0d") : failed to sync secret cache: timed out waiting for the condition Mar 20 09:00:28 crc kubenswrapper[4858]: E0320 09:00:28.345894 4858 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Mar 20 09:00:28 crc kubenswrapper[4858]: E0320 09:00:28.346002 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/960976f4-1bea-423a-b4fc-09b08a60ba0d-config podName:960976f4-1bea-423a-b4fc-09b08a60ba0d nodeName:}" failed. No retries permitted until 2026-03-20 09:00:28.845967669 +0000 UTC m=+210.166385916 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/960976f4-1bea-423a-b4fc-09b08a60ba0d-config") pod "service-ca-operator-777779d784-fzvwc" (UID: "960976f4-1bea-423a-b4fc-09b08a60ba0d") : failed to sync configmap cache: timed out waiting for the condition Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.358813 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.377426 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.396255 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.416759 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.436598 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.456342 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.460531 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-c5kzc"] Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.477708 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hsp4b"] Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.479419 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: W0320 09:00:28.488543 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod039cac36_f4ed_4282_aa07_ee40ad00df93.slice/crio-1aac04c6846e72c5160319abc1239d2160de18d4efacf3b517ec441ee0b0f1af WatchSource:0}: Error finding container 1aac04c6846e72c5160319abc1239d2160de18d4efacf3b517ec441ee0b0f1af: Status 404 returned error can't find the container with id 1aac04c6846e72c5160319abc1239d2160de18d4efacf3b517ec441ee0b0f1af Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.495617 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.516525 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.536604 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.556906 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.576701 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.597045 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.624493 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.637363 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.656200 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.676949 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.696303 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.716857 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.737090 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.756231 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.777012 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.797205 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.816778 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.837460 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.869973 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/960976f4-1bea-423a-b4fc-09b08a60ba0d-serving-cert\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.870220 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960976f4-1bea-423a-b4fc-09b08a60ba0d-config\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.871592 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/960976f4-1bea-423a-b4fc-09b08a60ba0d-config\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.873868 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/960976f4-1bea-423a-b4fc-09b08a60ba0d-serving-cert\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.876987 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.897844 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.916982 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.941145 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.957503 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 09:00:28 crc kubenswrapper[4858]: I0320 09:00:28.977407 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.019673 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sg8h\" (UniqueName: \"kubernetes.io/projected/e3892972-3419-4c52-bb2f-993e4a19d813-kube-api-access-6sg8h\") pod \"dns-operator-744455d44c-jsp8n\" (UID: \"e3892972-3419-4c52-bb2f-993e4a19d813\") " pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.038217 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghvr5\" (UniqueName: \"kubernetes.io/projected/1f93ac17-3d80-426f-9d3e-8d09ee8f84e6-kube-api-access-ghvr5\") pod \"authentication-operator-69f744f599-5grwr\" (UID: \"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.057138 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4mgr\" (UniqueName: \"kubernetes.io/projected/3c53fc26-4e6d-4d8f-bb46-59987bcc746f-kube-api-access-m4mgr\") pod \"machine-api-operator-5694c8668f-vw55r\" (UID: \"3c53fc26-4e6d-4d8f-bb46-59987bcc746f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.075525 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcn9q\" (UniqueName: \"kubernetes.io/projected/e07edf68-41a8-4175-adc0-163e46620ab4-kube-api-access-qcn9q\") pod \"downloads-7954f5f757-j6mmm\" (UID: \"e07edf68-41a8-4175-adc0-163e46620ab4\") " pod="openshift-console/downloads-7954f5f757-j6mmm" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.094651 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shhz7\" (UniqueName: \"kubernetes.io/projected/3d7f0079-7bd1-40b1-ba84-855d45b00dc0-kube-api-access-shhz7\") pod \"console-operator-58897d9998-cwnt6\" (UID: \"3d7f0079-7bd1-40b1-ba84-855d45b00dc0\") " pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.109406 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.118433 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxk6j\" (UniqueName: \"kubernetes.io/projected/055f16c2-9ca1-4078-82b9-48aa9a4399ad-kube-api-access-lxk6j\") pod \"apiserver-7bbb656c7d-vfg5z\" (UID: \"055f16c2-9ca1-4078-82b9-48aa9a4399ad\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.131741 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.137787 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vl4n\" (UniqueName: \"kubernetes.io/projected/830fcf94-999e-4859-a62e-f317fc53eaf6-kube-api-access-9vl4n\") pod \"router-default-5444994796-vtwn4\" (UID: \"830fcf94-999e-4859-a62e-f317fc53eaf6\") " pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.141995 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-j6mmm" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.157418 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85htn\" (UniqueName: \"kubernetes.io/projected/31262c04-c5c9-4b06-afd8-f005d271819a-kube-api-access-85htn\") pod \"machine-approver-56656f9798-pbg4m\" (UID: \"31262c04-c5c9-4b06-afd8-f005d271819a\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.174997 4858 request.go:700] Waited for 1.939963387s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.175751 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct5l6\" (UniqueName: \"kubernetes.io/projected/59f9f7fc-da90-46ec-b360-1eee512a4416-kube-api-access-ct5l6\") pod \"openshift-controller-manager-operator-756b6f6bc6-zdvzk\" (UID: \"59f9f7fc-da90-46ec-b360-1eee512a4416\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.194462 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v727w\" (UniqueName: \"kubernetes.io/projected/3b3bd918-04b9-4371-933a-609e9add5512-kube-api-access-v727w\") pod \"openshift-apiserver-operator-796bbdcf4f-bkvnd\" (UID: \"3b3bd918-04b9-4371-933a-609e9add5512\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.194892 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.196597 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.232877 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.236374 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tszd\" (UniqueName: \"kubernetes.io/projected/effb9468-b572-4eb1-84df-15e7b0201dbf-kube-api-access-7tszd\") pod \"route-controller-manager-6576b87f9c-sg9cs\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.256955 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.258426 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdvl5\" (UniqueName: \"kubernetes.io/projected/91be84d3-8196-44bb-8a88-e9e6548377a1-kube-api-access-qdvl5\") pod \"console-f9d7485db-wr84h\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.276872 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.281680 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.295994 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" event={"ID":"31262c04-c5c9-4b06-afd8-f005d271819a","Type":"ContainerStarted","Data":"d27de740b5ec7322f7b6679f6786b91456ca85370415f6bcf56d143d1ce3ea08"} Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.298007 4858 generic.go:334] "Generic (PLEG): container finished" podID="84a3f478-6ec6-4ef6-983e-acdeeb3b475c" containerID="47a2f8619644faa4464b9845245c220edcfaf613abc136a9a8bb63b443f123b1" exitCode=0 Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.298090 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" event={"ID":"84a3f478-6ec6-4ef6-983e-acdeeb3b475c","Type":"ContainerDied","Data":"47a2f8619644faa4464b9845245c220edcfaf613abc136a9a8bb63b443f123b1"} Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.298137 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" event={"ID":"84a3f478-6ec6-4ef6-983e-acdeeb3b475c","Type":"ContainerStarted","Data":"b79a9cf32e38678f5c0cfdbb4d7fcb8507b6ab9062b4fa5f6b771b8a25d79eff"} Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.308859 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" event={"ID":"039cac36-f4ed-4282-aa07-ee40ad00df93","Type":"ContainerStarted","Data":"6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259"} Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.308908 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" event={"ID":"039cac36-f4ed-4282-aa07-ee40ad00df93","Type":"ContainerStarted","Data":"1aac04c6846e72c5160319abc1239d2160de18d4efacf3b517ec441ee0b0f1af"} Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.309853 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.310306 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.312675 4858 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-hsp4b container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.312717 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" podUID="039cac36-f4ed-4282-aa07-ee40ad00df93" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.317797 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8tvn\" (UniqueName: \"kubernetes.io/projected/c245d181-9680-448c-a0c6-32f5d54811f7-kube-api-access-m8tvn\") pod \"cluster-samples-operator-665b6dd947-xm7ws\" (UID: \"c245d181-9680-448c-a0c6-32f5d54811f7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.323676 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.332031 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.338897 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxr8p\" (UniqueName: \"kubernetes.io/projected/990e152e-f7bc-4811-bc9a-6954a09b166a-kube-api-access-gxr8p\") pod \"openshift-config-operator-7777fb866f-xbjjz\" (UID: \"990e152e-f7bc-4811-bc9a-6954a09b166a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.354626 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w92k6\" (UniqueName: \"kubernetes.io/projected/99346cc6-9090-4e06-beb0-d64a92bd2813-kube-api-access-w92k6\") pod \"etcd-operator-b45778765-468wl\" (UID: \"99346cc6-9090-4e06-beb0-d64a92bd2813\") " pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.375244 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtww9\" (UniqueName: \"kubernetes.io/projected/a9f367b2-d0b3-4a80-933f-68bf11e63791-kube-api-access-vtww9\") pod \"oauth-openshift-558db77b4-2c66r\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.375961 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.376305 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-jsp8n"] Mar 20 09:00:29 crc kubenswrapper[4858]: W0320 09:00:29.393817 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod830fcf94_999e_4859_a62e_f317fc53eaf6.slice/crio-687bd8aa4b32b285fa3e9e9b8e5bcfbf94f6348f0287543767c098c6d4d3350a WatchSource:0}: Error finding container 687bd8aa4b32b285fa3e9e9b8e5bcfbf94f6348f0287543767c098c6d4d3350a: Status 404 returned error can't find the container with id 687bd8aa4b32b285fa3e9e9b8e5bcfbf94f6348f0287543767c098c6d4d3350a Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.401764 4858 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.406596 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5grwr"] Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.418741 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.429997 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-j6mmm"] Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.434090 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.436358 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.451857 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.457813 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.460139 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.484565 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.485674 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.517625 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6scs\" (UniqueName: \"kubernetes.io/projected/74fe10ec-a162-4c93-b2d3-1a80745e7fcc-kube-api-access-v6scs\") pod \"auto-csr-approver-29566620-gz4tx\" (UID: \"74fe10ec-a162-4c93-b2d3-1a80745e7fcc\") " pod="openshift-infra/auto-csr-approver-29566620-gz4tx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.553008 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q65g9\" (UniqueName: \"kubernetes.io/projected/e2f99651-1c5a-4f42-a46e-af580ec9b4eb-kube-api-access-q65g9\") pod \"machine-config-operator-74547568cd-dcfwn\" (UID: \"e2f99651-1c5a-4f42-a46e-af580ec9b4eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.561129 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.567123 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.583273 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swq66\" (UniqueName: \"kubernetes.io/projected/69e5a3c1-6c66-42f5-a122-63c4d2838aca-kube-api-access-swq66\") pod \"catalog-operator-68c6474976-rhqsx\" (UID: \"69e5a3c1-6c66-42f5-a122-63c4d2838aca\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.598131 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.612739 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.641142 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd"] Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.650624 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwfzf\" (UniqueName: \"kubernetes.io/projected/960976f4-1bea-423a-b4fc-09b08a60ba0d-kube-api-access-dwfzf\") pod \"service-ca-operator-777779d784-fzvwc\" (UID: \"960976f4-1bea-423a-b4fc-09b08a60ba0d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.663064 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689159 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689201 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6287e483-4f8f-4be6-840c-a42d3420d3a5-srv-cert\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689236 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bc9663-4753-4ad1-b0f4-2414dc389098-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689300 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnbmb\" (UniqueName: \"kubernetes.io/projected/16f8cb13-edc5-403c-bc8c-2fcd585139b0-kube-api-access-mnbmb\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689362 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-config\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689382 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/631923d2-1540-4a96-889a-d6f39d28ef1b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689400 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/631923d2-1540-4a96-889a-d6f39d28ef1b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689435 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/563363e9-546c-4754-93b8-c274c58779b0-webhook-cert\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689462 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689502 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-bound-sa-token\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689520 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16f8cb13-edc5-403c-bc8c-2fcd585139b0-metrics-tls\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689541 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e587c0c9-bf99-4f08-96de-e3a386af8b8a-signing-key\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689560 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcbxz\" (UniqueName: \"kubernetes.io/projected/2cc056e7-0895-499c-acb5-0c82b7a8b900-kube-api-access-qcbxz\") pod \"multus-admission-controller-857f4d67dd-h64jx\" (UID: \"2cc056e7-0895-499c-acb5-0c82b7a8b900\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689582 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2fae0532-0891-4c58-abec-e48437904f40-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689620 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghzv7\" (UniqueName: \"kubernetes.io/projected/563363e9-546c-4754-93b8-c274c58779b0-kube-api-access-ghzv7\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689659 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc30d82f-f02c-41c4-9c6e-94e663fa8712-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689685 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a91bc770-3810-4e8e-ac89-4b321be44b3c-proxy-tls\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689723 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gjsq\" (UniqueName: \"kubernetes.io/projected/6db678cf-767f-4339-90db-09aa1fe57983-kube-api-access-4gjsq\") pod \"package-server-manager-789f6589d5-qgkz2\" (UID: \"6db678cf-767f-4339-90db-09aa1fe57983\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689738 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbdqn\" (UniqueName: \"kubernetes.io/projected/631923d2-1540-4a96-889a-d6f39d28ef1b-kube-api-access-mbdqn\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689765 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16f8cb13-edc5-403c-bc8c-2fcd585139b0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689781 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6db678cf-767f-4339-90db-09aa1fe57983-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qgkz2\" (UID: \"6db678cf-767f-4339-90db-09aa1fe57983\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689811 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8f37a14-0144-4154-a087-126fde1633eb-config-volume\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689829 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-registry-certificates\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689844 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689860 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc30d82f-f02c-41c4-9c6e-94e663fa8712-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689878 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689897 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnh8z\" (UniqueName: \"kubernetes.io/projected/b51deb4e-ca50-41d4-8b00-bb996f8e7782-kube-api-access-fnh8z\") pod \"control-plane-machine-set-operator-78cbb6b69f-msrvs\" (UID: \"b51deb4e-ca50-41d4-8b00-bb996f8e7782\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.689940 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snsr6\" (UniqueName: \"kubernetes.io/projected/dc30d82f-f02c-41c4-9c6e-94e663fa8712-kube-api-access-snsr6\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690124 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fae0532-0891-4c58-abec-e48437904f40-config\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690144 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfh7c\" (UniqueName: \"kubernetes.io/projected/afdaaa04-0b97-4c8c-8699-a620209b9202-kube-api-access-qfh7c\") pod \"migrator-59844c95c7-z9dfb\" (UID: \"afdaaa04-0b97-4c8c-8699-a620209b9202\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690163 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6287e483-4f8f-4be6-840c-a42d3420d3a5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690205 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e587c0c9-bf99-4f08-96de-e3a386af8b8a-signing-cabundle\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690231 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x79ww\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-kube-api-access-x79ww\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690256 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8f37a14-0144-4154-a087-126fde1633eb-secret-volume\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690286 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-registry-tls\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690457 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2cc056e7-0895-499c-acb5-0c82b7a8b900-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h64jx\" (UID: \"2cc056e7-0895-499c-acb5-0c82b7a8b900\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690513 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-trusted-ca\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690531 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/563363e9-546c-4754-93b8-c274c58779b0-apiservice-cert\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690548 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxg4l\" (UniqueName: \"kubernetes.io/projected/d8f37a14-0144-4154-a087-126fde1633eb-kube-api-access-pxg4l\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690567 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a91bc770-3810-4e8e-ac89-4b321be44b3c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690605 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5lkn\" (UniqueName: \"kubernetes.io/projected/a91bc770-3810-4e8e-ac89-4b321be44b3c-kube-api-access-t5lkn\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690624 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fae0532-0891-4c58-abec-e48437904f40-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690656 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2e4d497-a390-4102-961e-8334641b8867-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690673 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2e4d497-a390-4102-961e-8334641b8867-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690731 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48bc9663-4753-4ad1-b0f4-2414dc389098-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690760 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48bc9663-4753-4ad1-b0f4-2414dc389098-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690809 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b51deb4e-ca50-41d4-8b00-bb996f8e7782-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-msrvs\" (UID: \"b51deb4e-ca50-41d4-8b00-bb996f8e7782\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690831 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/631923d2-1540-4a96-889a-d6f39d28ef1b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690854 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/563363e9-546c-4754-93b8-c274c58779b0-tmpfs\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690917 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtb5j\" (UniqueName: \"kubernetes.io/projected/e587c0c9-bf99-4f08-96de-e3a386af8b8a-kube-api-access-xtb5j\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.690984 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.691026 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldb69\" (UniqueName: \"kubernetes.io/projected/80c13002-5ff3-43ea-be87-e1b2ecf4431a-kube-api-access-ldb69\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.691045 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d74bj\" (UniqueName: \"kubernetes.io/projected/6287e483-4f8f-4be6-840c-a42d3420d3a5-kube-api-access-d74bj\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.691064 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16f8cb13-edc5-403c-bc8c-2fcd585139b0-trusted-ca\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: E0320 09:00:29.698870 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.198852662 +0000 UTC m=+211.519270859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.744071 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vw55r"] Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.774111 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z"] Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.779838 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-cwnt6"] Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792514 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792753 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-bound-sa-token\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792784 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16f8cb13-edc5-403c-bc8c-2fcd585139b0-metrics-tls\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792803 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e587c0c9-bf99-4f08-96de-e3a386af8b8a-signing-key\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792830 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcbxz\" (UniqueName: \"kubernetes.io/projected/2cc056e7-0895-499c-acb5-0c82b7a8b900-kube-api-access-qcbxz\") pod \"multus-admission-controller-857f4d67dd-h64jx\" (UID: \"2cc056e7-0895-499c-acb5-0c82b7a8b900\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792847 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2fae0532-0891-4c58-abec-e48437904f40-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792867 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-mountpoint-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792887 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghzv7\" (UniqueName: \"kubernetes.io/projected/563363e9-546c-4754-93b8-c274c58779b0-kube-api-access-ghzv7\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792915 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc30d82f-f02c-41c4-9c6e-94e663fa8712-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792932 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a91bc770-3810-4e8e-ac89-4b321be44b3c-proxy-tls\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792964 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gjsq\" (UniqueName: \"kubernetes.io/projected/6db678cf-767f-4339-90db-09aa1fe57983-kube-api-access-4gjsq\") pod \"package-server-manager-789f6589d5-qgkz2\" (UID: \"6db678cf-767f-4339-90db-09aa1fe57983\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792981 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbdqn\" (UniqueName: \"kubernetes.io/projected/631923d2-1540-4a96-889a-d6f39d28ef1b-kube-api-access-mbdqn\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.792999 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16f8cb13-edc5-403c-bc8c-2fcd585139b0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793015 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-csi-data-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793032 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6db678cf-767f-4339-90db-09aa1fe57983-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qgkz2\" (UID: \"6db678cf-767f-4339-90db-09aa1fe57983\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793060 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8f37a14-0144-4154-a087-126fde1633eb-config-volume\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793077 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-registry-certificates\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793095 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793122 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc30d82f-f02c-41c4-9c6e-94e663fa8712-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793138 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793162 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnh8z\" (UniqueName: \"kubernetes.io/projected/b51deb4e-ca50-41d4-8b00-bb996f8e7782-kube-api-access-fnh8z\") pod \"control-plane-machine-set-operator-78cbb6b69f-msrvs\" (UID: \"b51deb4e-ca50-41d4-8b00-bb996f8e7782\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793200 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-socket-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793229 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snsr6\" (UniqueName: \"kubernetes.io/projected/dc30d82f-f02c-41c4-9c6e-94e663fa8712-kube-api-access-snsr6\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793279 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrfdx\" (UniqueName: \"kubernetes.io/projected/978fa44d-6fb3-4775-ad8f-d81a582b521e-kube-api-access-vrfdx\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793346 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fae0532-0891-4c58-abec-e48437904f40-config\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793365 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfh7c\" (UniqueName: \"kubernetes.io/projected/afdaaa04-0b97-4c8c-8699-a620209b9202-kube-api-access-qfh7c\") pod \"migrator-59844c95c7-z9dfb\" (UID: \"afdaaa04-0b97-4c8c-8699-a620209b9202\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793384 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqqzb\" (UniqueName: \"kubernetes.io/projected/4c2dfb4c-a6cb-4736-a668-020e93ffe5f0-kube-api-access-pqqzb\") pod \"ingress-canary-b58fr\" (UID: \"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0\") " pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793412 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6287e483-4f8f-4be6-840c-a42d3420d3a5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793427 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-certs\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793451 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/95a6716d-90d6-4984-a85c-eab9192d4d75-metrics-tls\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793468 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e587c0c9-bf99-4f08-96de-e3a386af8b8a-signing-cabundle\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793493 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x79ww\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-kube-api-access-x79ww\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793510 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8f37a14-0144-4154-a087-126fde1633eb-secret-volume\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793536 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-registry-tls\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793562 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phkr4\" (UniqueName: \"kubernetes.io/projected/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-kube-api-access-phkr4\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793579 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2cc056e7-0895-499c-acb5-0c82b7a8b900-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h64jx\" (UID: \"2cc056e7-0895-499c-acb5-0c82b7a8b900\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793615 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-trusted-ca\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793630 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/563363e9-546c-4754-93b8-c274c58779b0-apiservice-cert\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793646 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxg4l\" (UniqueName: \"kubernetes.io/projected/d8f37a14-0144-4154-a087-126fde1633eb-kube-api-access-pxg4l\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793662 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a91bc770-3810-4e8e-ac89-4b321be44b3c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793685 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5lkn\" (UniqueName: \"kubernetes.io/projected/a91bc770-3810-4e8e-ac89-4b321be44b3c-kube-api-access-t5lkn\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793717 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fae0532-0891-4c58-abec-e48437904f40-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793734 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2e4d497-a390-4102-961e-8334641b8867-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793755 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2e4d497-a390-4102-961e-8334641b8867-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793776 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95a6716d-90d6-4984-a85c-eab9192d4d75-config-volume\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793798 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48bc9663-4753-4ad1-b0f4-2414dc389098-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793813 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-node-bootstrap-token\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793839 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48bc9663-4753-4ad1-b0f4-2414dc389098-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793887 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b51deb4e-ca50-41d4-8b00-bb996f8e7782-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-msrvs\" (UID: \"b51deb4e-ca50-41d4-8b00-bb996f8e7782\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793906 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/631923d2-1540-4a96-889a-d6f39d28ef1b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793921 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/563363e9-546c-4754-93b8-c274c58779b0-tmpfs\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793958 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7s7l\" (UniqueName: \"kubernetes.io/projected/95a6716d-90d6-4984-a85c-eab9192d4d75-kube-api-access-k7s7l\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.793974 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4c2dfb4c-a6cb-4736-a668-020e93ffe5f0-cert\") pod \"ingress-canary-b58fr\" (UID: \"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0\") " pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794010 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtb5j\" (UniqueName: \"kubernetes.io/projected/e587c0c9-bf99-4f08-96de-e3a386af8b8a-kube-api-access-xtb5j\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794025 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-plugins-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794076 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldb69\" (UniqueName: \"kubernetes.io/projected/80c13002-5ff3-43ea-be87-e1b2ecf4431a-kube-api-access-ldb69\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794095 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d74bj\" (UniqueName: \"kubernetes.io/projected/6287e483-4f8f-4be6-840c-a42d3420d3a5-kube-api-access-d74bj\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794113 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16f8cb13-edc5-403c-bc8c-2fcd585139b0-trusted-ca\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794129 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794148 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6287e483-4f8f-4be6-840c-a42d3420d3a5-srv-cert\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794164 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bc9663-4753-4ad1-b0f4-2414dc389098-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794188 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnbmb\" (UniqueName: \"kubernetes.io/projected/16f8cb13-edc5-403c-bc8c-2fcd585139b0-kube-api-access-mnbmb\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794204 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-registration-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794230 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-config\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794246 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/631923d2-1540-4a96-889a-d6f39d28ef1b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794262 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/631923d2-1540-4a96-889a-d6f39d28ef1b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794295 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/563363e9-546c-4754-93b8-c274c58779b0-webhook-cert\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.794336 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.801089 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2fae0532-0891-4c58-abec-e48437904f40-config\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: E0320 09:00:29.801356 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.301336879 +0000 UTC m=+211.621755076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.802550 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.804341 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc30d82f-f02c-41c4-9c6e-94e663fa8712-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.804992 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8f37a14-0144-4154-a087-126fde1633eb-config-volume\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.807297 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/563363e9-546c-4754-93b8-c274c58779b0-tmpfs\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.808449 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16f8cb13-edc5-403c-bc8c-2fcd585139b0-trusted-ca\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.808958 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-trusted-ca\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.809451 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-config\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.810745 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e587c0c9-bf99-4f08-96de-e3a386af8b8a-signing-cabundle\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.810834 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2e4d497-a390-4102-961e-8334641b8867-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.812689 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a91bc770-3810-4e8e-ac89-4b321be44b3c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.813586 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bc9663-4753-4ad1-b0f4-2414dc389098-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.813769 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/631923d2-1540-4a96-889a-d6f39d28ef1b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.814108 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/b51deb4e-ca50-41d4-8b00-bb996f8e7782-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-msrvs\" (UID: \"b51deb4e-ca50-41d4-8b00-bb996f8e7782\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.816111 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.816209 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.816619 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/631923d2-1540-4a96-889a-d6f39d28ef1b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.816744 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6287e483-4f8f-4be6-840c-a42d3420d3a5-srv-cert\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.817477 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-registry-certificates\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.817945 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6db678cf-767f-4339-90db-09aa1fe57983-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-qgkz2\" (UID: \"6db678cf-767f-4339-90db-09aa1fe57983\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.818213 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.818748 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/16f8cb13-edc5-403c-bc8c-2fcd585139b0-metrics-tls\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.819500 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48bc9663-4753-4ad1-b0f4-2414dc389098-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.819666 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8f37a14-0144-4154-a087-126fde1633eb-secret-volume\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.820382 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e587c0c9-bf99-4f08-96de-e3a386af8b8a-signing-key\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.820435 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/2cc056e7-0895-499c-acb5-0c82b7a8b900-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-h64jx\" (UID: \"2cc056e7-0895-499c-acb5-0c82b7a8b900\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.821124 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-registry-tls\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.823012 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6287e483-4f8f-4be6-840c-a42d3420d3a5-profile-collector-cert\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.824831 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2e4d497-a390-4102-961e-8334641b8867-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.825584 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/563363e9-546c-4754-93b8-c274c58779b0-apiservice-cert\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.828222 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc30d82f-f02c-41c4-9c6e-94e663fa8712-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.828659 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2fae0532-0891-4c58-abec-e48437904f40-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.830178 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/563363e9-546c-4754-93b8-c274c58779b0-webhook-cert\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.830352 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a91bc770-3810-4e8e-ac89-4b321be44b3c-proxy-tls\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.833264 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfh7c\" (UniqueName: \"kubernetes.io/projected/afdaaa04-0b97-4c8c-8699-a620209b9202-kube-api-access-qfh7c\") pod \"migrator-59844c95c7-z9dfb\" (UID: \"afdaaa04-0b97-4c8c-8699-a620209b9202\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.875594 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxg4l\" (UniqueName: \"kubernetes.io/projected/d8f37a14-0144-4154-a087-126fde1633eb-kube-api-access-pxg4l\") pod \"collect-profiles-29566620-tjqsd\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.876046 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x79ww\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-kube-api-access-x79ww\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.886452 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.896365 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-node-bootstrap-token\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.896718 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7s7l\" (UniqueName: \"kubernetes.io/projected/95a6716d-90d6-4984-a85c-eab9192d4d75-kube-api-access-k7s7l\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897063 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4c2dfb4c-a6cb-4736-a668-020e93ffe5f0-cert\") pod \"ingress-canary-b58fr\" (UID: \"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0\") " pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897161 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897304 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-plugins-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897424 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-registration-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897543 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-mountpoint-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897653 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-csi-data-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897754 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-socket-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.897854 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrfdx\" (UniqueName: \"kubernetes.io/projected/978fa44d-6fb3-4775-ad8f-d81a582b521e-kube-api-access-vrfdx\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.898146 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-registration-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.898228 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-csi-data-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.898257 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-mountpoint-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: E0320 09:00:29.898284 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.398267568 +0000 UTC m=+211.718685765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.899123 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqqzb\" (UniqueName: \"kubernetes.io/projected/4c2dfb4c-a6cb-4736-a668-020e93ffe5f0-kube-api-access-pqqzb\") pod \"ingress-canary-b58fr\" (UID: \"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0\") " pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.899226 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-certs\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.899333 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/95a6716d-90d6-4984-a85c-eab9192d4d75-metrics-tls\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.899442 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phkr4\" (UniqueName: \"kubernetes.io/projected/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-kube-api-access-phkr4\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.898298 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-plugins-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.898294 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/978fa44d-6fb3-4775-ad8f-d81a582b521e-socket-dir\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.901900 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-node-bootstrap-token\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.902537 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghzv7\" (UniqueName: \"kubernetes.io/projected/563363e9-546c-4754-93b8-c274c58779b0-kube-api-access-ghzv7\") pod \"packageserver-d55dfcdfc-nvr22\" (UID: \"563363e9-546c-4754-93b8-c274c58779b0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.903083 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95a6716d-90d6-4984-a85c-eab9192d4d75-config-volume\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.903787 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-certs\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.906566 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4c2dfb4c-a6cb-4736-a668-020e93ffe5f0-cert\") pod \"ingress-canary-b58fr\" (UID: \"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0\") " pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.916179 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.919167 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5lkn\" (UniqueName: \"kubernetes.io/projected/a91bc770-3810-4e8e-ac89-4b321be44b3c-kube-api-access-t5lkn\") pod \"machine-config-controller-84d6567774-z9scw\" (UID: \"a91bc770-3810-4e8e-ac89-4b321be44b3c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.922920 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/95a6716d-90d6-4984-a85c-eab9192d4d75-metrics-tls\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.930924 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.934923 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/95a6716d-90d6-4984-a85c-eab9192d4d75-config-volume\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.939483 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-bound-sa-token\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.972939 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/48bc9663-4753-4ad1-b0f4-2414dc389098-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bv8pj\" (UID: \"48bc9663-4753-4ad1-b0f4-2414dc389098\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:29 crc kubenswrapper[4858]: I0320 09:00:29.988876 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcbxz\" (UniqueName: \"kubernetes.io/projected/2cc056e7-0895-499c-acb5-0c82b7a8b900-kube-api-access-qcbxz\") pod \"multus-admission-controller-857f4d67dd-h64jx\" (UID: \"2cc056e7-0895-499c-acb5-0c82b7a8b900\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:29.999696 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-wr84h"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.003962 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2fae0532-0891-4c58-abec-e48437904f40-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-s487j\" (UID: \"2fae0532-0891-4c58-abec-e48437904f40\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.004997 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.005469 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.505452202 +0000 UTC m=+211.825870399 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.036048 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rndqc\" (UID: \"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.059389 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snsr6\" (UniqueName: \"kubernetes.io/projected/dc30d82f-f02c-41c4-9c6e-94e663fa8712-kube-api-access-snsr6\") pod \"kube-storage-version-migrator-operator-b67b599dd-qw67f\" (UID: \"dc30d82f-f02c-41c4-9c6e-94e663fa8712\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.067395 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnh8z\" (UniqueName: \"kubernetes.io/projected/b51deb4e-ca50-41d4-8b00-bb996f8e7782-kube-api-access-fnh8z\") pod \"control-plane-machine-set-operator-78cbb6b69f-msrvs\" (UID: \"b51deb4e-ca50-41d4-8b00-bb996f8e7782\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.079556 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbdqn\" (UniqueName: \"kubernetes.io/projected/631923d2-1540-4a96-889a-d6f39d28ef1b-kube-api-access-mbdqn\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:30 crc kubenswrapper[4858]: W0320 09:00:30.100461 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91be84d3_8196_44bb_8a88_e9e6548377a1.slice/crio-d1092aa8af91825451b427caff92dbdeacfc64fbef2476b11042ed5f53e4e62b WatchSource:0}: Error finding container d1092aa8af91825451b427caff92dbdeacfc64fbef2476b11042ed5f53e4e62b: Status 404 returned error can't find the container with id d1092aa8af91825451b427caff92dbdeacfc64fbef2476b11042ed5f53e4e62b Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.106331 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gjsq\" (UniqueName: \"kubernetes.io/projected/6db678cf-767f-4339-90db-09aa1fe57983-kube-api-access-4gjsq\") pod \"package-server-manager-789f6589d5-qgkz2\" (UID: \"6db678cf-767f-4339-90db-09aa1fe57983\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.119079 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.119480 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.619467271 +0000 UTC m=+211.939885468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.127080 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16f8cb13-edc5-403c-bc8c-2fcd585139b0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.155120 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.156507 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.166137 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.169182 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldb69\" (UniqueName: \"kubernetes.io/projected/80c13002-5ff3-43ea-be87-e1b2ecf4431a-kube-api-access-ldb69\") pod \"marketplace-operator-79b997595-4s5wf\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.171271 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.192895 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtb5j\" (UniqueName: \"kubernetes.io/projected/e587c0c9-bf99-4f08-96de-e3a386af8b8a-kube-api-access-xtb5j\") pod \"service-ca-9c57cc56f-9rv48\" (UID: \"e587c0c9-bf99-4f08-96de-e3a386af8b8a\") " pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.193202 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.194093 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.197297 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d74bj\" (UniqueName: \"kubernetes.io/projected/6287e483-4f8f-4be6-840c-a42d3420d3a5-kube-api-access-d74bj\") pod \"olm-operator-6b444d44fb-vk2kc\" (UID: \"6287e483-4f8f-4be6-840c-a42d3420d3a5\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.201905 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.208644 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.208777 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnbmb\" (UniqueName: \"kubernetes.io/projected/16f8cb13-edc5-403c-bc8c-2fcd585139b0-kube-api-access-mnbmb\") pod \"ingress-operator-5b745b69d9-6mj9n\" (UID: \"16f8cb13-edc5-403c-bc8c-2fcd585139b0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.218041 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/631923d2-1540-4a96-889a-d6f39d28ef1b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-rh86k\" (UID: \"631923d2-1540-4a96-889a-d6f39d28ef1b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.220877 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.221456 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.721412326 +0000 UTC m=+212.041830523 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.224439 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.240308 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.253915 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.253954 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-468wl"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.258952 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7s7l\" (UniqueName: \"kubernetes.io/projected/95a6716d-90d6-4984-a85c-eab9192d4d75-kube-api-access-k7s7l\") pod \"dns-default-c47vz\" (UID: \"95a6716d-90d6-4984-a85c-eab9192d4d75\") " pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.305042 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.319457 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.322904 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.323427 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.8233943 +0000 UTC m=+212.143812487 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.340078 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqqzb\" (UniqueName: \"kubernetes.io/projected/4c2dfb4c-a6cb-4736-a668-020e93ffe5f0-kube-api-access-pqqzb\") pod \"ingress-canary-b58fr\" (UID: \"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0\") " pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.352399 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phkr4\" (UniqueName: \"kubernetes.io/projected/8ab7cf82-a1db-4139-82eb-9d05c0df6a50-kube-api-access-phkr4\") pod \"machine-config-server-rzvjq\" (UID: \"8ab7cf82-a1db-4139-82eb-9d05c0df6a50\") " pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.362094 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.362603 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" event={"ID":"3b3bd918-04b9-4371-933a-609e9add5512","Type":"ContainerStarted","Data":"301d64db542c25c7e04a5305e071c0b8d4f0f7358077bf129424fcc015666d51"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.362628 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" event={"ID":"99346cc6-9090-4e06-beb0-d64a92bd2813","Type":"ContainerStarted","Data":"50eae976ccaa13b68ba4ea6e6e663e80dde8dd5739237266e3e804753cea4ebd"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.403661 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrfdx\" (UniqueName: \"kubernetes.io/projected/978fa44d-6fb3-4775-ad8f-d81a582b521e-kube-api-access-vrfdx\") pod \"csi-hostpathplugin-tbsv5\" (UID: \"978fa44d-6fb3-4775-ad8f-d81a582b521e\") " pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.409351 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.423934 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.428666 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.428924 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.430467 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:30.930444031 +0000 UTC m=+212.250862228 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.432550 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2c66r"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.432751 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.437480 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" event={"ID":"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6","Type":"ContainerStarted","Data":"cfe061b69b26052f9545226f607bf238f3caf292bba4b9af926d5fa93cb88508"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.437542 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" event={"ID":"1f93ac17-3d80-426f-9d3e-8d09ee8f84e6","Type":"ContainerStarted","Data":"f37ad03e5ce24d91d8b28431b3edbf79f4e1b7eda0e97d8f2daa99e1a37869df"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.474962 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" event={"ID":"e3892972-3419-4c52-bb2f-993e4a19d813","Type":"ContainerStarted","Data":"e40ffc3e088b1f3dcb98bc7111e03d6756a331c61af1771e372cf43b49c1ace7"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.475037 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" event={"ID":"e3892972-3419-4c52-bb2f-993e4a19d813","Type":"ContainerStarted","Data":"ad4e288f6e9c0f1ca485af9026f26ac8d60370ca5a7ffb93a982fcc43d412b43"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.477658 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" event={"ID":"59f9f7fc-da90-46ec-b360-1eee512a4416","Type":"ContainerStarted","Data":"4e29c4f4eb62551cdb559589dda1d05a49de5c2ccc92b32c50bccd88c295f695"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.529217 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.530421 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.03040911 +0000 UTC m=+212.350827307 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.531753 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.546059 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" event={"ID":"84a3f478-6ec6-4ef6-983e-acdeeb3b475c","Type":"ContainerStarted","Data":"02387e25d25ca96006465b07327e45c7a91611c4ed48a8a4ebe786be586a909a"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.561165 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" event={"ID":"3d7f0079-7bd1-40b1-ba84-855d45b00dc0","Type":"ContainerStarted","Data":"80c83992017e546c0603ee603086bdca765c1b6789886110dd58cf96e5f9b049"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.579884 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-b58fr" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.587749 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rzvjq" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.590645 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-j6mmm" event={"ID":"e07edf68-41a8-4175-adc0-163e46620ab4","Type":"ContainerStarted","Data":"66398f115820f09987f586ed105bafab1638dd2c8de815ad76df2187d6ae0824"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.590690 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-j6mmm" event={"ID":"e07edf68-41a8-4175-adc0-163e46620ab4","Type":"ContainerStarted","Data":"045fe05c663363173acbda5715ab9f63a4c12e33946e07d0c7d531e617a82eae"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.597718 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.598150 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-j6mmm" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.599546 4858 patch_prober.go:28] interesting pod/downloads-7954f5f757-j6mmm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.599583 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-j6mmm" podUID="e07edf68-41a8-4175-adc0-163e46620ab4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.600029 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-vtwn4" event={"ID":"830fcf94-999e-4859-a62e-f317fc53eaf6","Type":"ContainerStarted","Data":"704665e80848bd9c697adb9be425d982fb3608e6ab2a3bd9a45ad1f90df63ccc"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.600085 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-vtwn4" event={"ID":"830fcf94-999e-4859-a62e-f317fc53eaf6","Type":"ContainerStarted","Data":"687bd8aa4b32b285fa3e9e9b8e5bcfbf94f6348f0287543767c098c6d4d3350a"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.603091 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz"] Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.604827 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wr84h" event={"ID":"91be84d3-8196-44bb-8a88-e9e6548377a1","Type":"ContainerStarted","Data":"d1092aa8af91825451b427caff92dbdeacfc64fbef2476b11042ed5f53e4e62b"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.606369 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" event={"ID":"3c53fc26-4e6d-4d8f-bb46-59987bcc746f","Type":"ContainerStarted","Data":"05faa35aaa9b8a4cd49799a44b80de734485648b8a5b0f4b5921a8bd994c4d42"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.608008 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" event={"ID":"055f16c2-9ca1-4078-82b9-48aa9a4399ad","Type":"ContainerStarted","Data":"ce22302493a0c6c8519b70f0f976553055a62170305bfcffebc712aa6866bdc4"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.613590 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" event={"ID":"31262c04-c5c9-4b06-afd8-f005d271819a","Type":"ContainerStarted","Data":"87999789ebc88fa42c9396ee32a268dfb92632ceea595f735016b771de5c6c7d"} Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.619392 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.631310 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.631912 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.131894533 +0000 UTC m=+212.452312730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.732942 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.736061 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.236033137 +0000 UTC m=+212.556451524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.834855 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.835698 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.335636179 +0000 UTC m=+212.656054376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:30 crc kubenswrapper[4858]: I0320 09:00:30.936435 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:30 crc kubenswrapper[4858]: E0320 09:00:30.937669 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.437640804 +0000 UTC m=+212.758059001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.037764 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.038355 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.538326639 +0000 UTC m=+212.858744836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.150200 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.151864 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.651837668 +0000 UTC m=+212.972255865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.153808 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.153982 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.154062 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.154261 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.166365 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.179011 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.203185 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.233751 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd"] Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.257172 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.257533 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.757496757 +0000 UTC m=+213.077914954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.257964 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.258559 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.75854686 +0000 UTC m=+213.078965057 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.263989 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.270185 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx"] Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.285603 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.293757 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.308352 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.336081 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.350730 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5grwr" podStartSLOduration=156.350678411 podStartE2EDuration="2m36.350678411s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:31.335928625 +0000 UTC m=+212.656346822" watchObservedRunningTime="2026-03-20 09:00:31.350678411 +0000 UTC m=+212.671096608" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.370039 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.370616 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.870584284 +0000 UTC m=+213.191002491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.451975 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-gz4tx"] Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.464522 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb"] Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.471547 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.472042 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:31.972025087 +0000 UTC m=+213.292443284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.574816 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.575328 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.075290671 +0000 UTC m=+213.395708868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.627739 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" event={"ID":"d8f37a14-0144-4154-a087-126fde1633eb","Type":"ContainerStarted","Data":"1f6ea64cacb834da961c70eae2de0ad2e965343c0d0bdb690ae7ded7e3fc0927"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.632036 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" event={"ID":"e2f99651-1c5a-4f42-a46e-af580ec9b4eb","Type":"ContainerStarted","Data":"51e2427e9433cbac6963fccd345e556b9a4c5907ac78419ee945dae358c410a7"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.643109 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" event={"ID":"e3892972-3419-4c52-bb2f-993e4a19d813","Type":"ContainerStarted","Data":"e4078dd21a1997f4888c2b312a1c21652c8b65b1763e3976d4604648742320aa"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.647609 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" event={"ID":"c245d181-9680-448c-a0c6-32f5d54811f7","Type":"ContainerStarted","Data":"256656c867be98d87101f73efe1ea6deb731e3af1e5da61b41784e0d3d0a97ff"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.649849 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw"] Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.662350 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-j6mmm" podStartSLOduration=156.662302575 podStartE2EDuration="2m36.662302575s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:31.653159627 +0000 UTC m=+212.973577824" watchObservedRunningTime="2026-03-20 09:00:31.662302575 +0000 UTC m=+212.982720762" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.664364 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" event={"ID":"69e5a3c1-6c66-42f5-a122-63c4d2838aca","Type":"ContainerStarted","Data":"efc68d10145c7ea321e5f805288cdcfcbfa5f60ece6f4d3317033711d52913ab"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.673596 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:31 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:31 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:31 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.673649 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.677160 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.677628 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.177610313 +0000 UTC m=+213.498028510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.681085 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" event={"ID":"3c53fc26-4e6d-4d8f-bb46-59987bcc746f","Type":"ContainerStarted","Data":"c0aff38c2120ade319c98c52f703b77c4e9da878d2dee2c96b11686c38f86117"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.688460 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc"] Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.688561 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" event={"ID":"990e152e-f7bc-4811-bc9a-6954a09b166a","Type":"ContainerStarted","Data":"486faf810c746685f1cf6d53e8db27b234eb1781f82ef61ad079c660eafa4633"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.695713 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" event={"ID":"3d7f0079-7bd1-40b1-ba84-855d45b00dc0","Type":"ContainerStarted","Data":"9da30c55380b17ddcd5b6b7f8a7203887e2be57deabe1267bf54790eed7c52be"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.696362 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.714843 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" event={"ID":"effb9468-b572-4eb1-84df-15e7b0201dbf","Type":"ContainerStarted","Data":"81daaedce65b3bc7d29a1d604b0acf510de0aebdcf9269fd1d4f8e329dad717c"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.715019 4858 patch_prober.go:28] interesting pod/console-operator-58897d9998-cwnt6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.715107 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" podUID="3d7f0079-7bd1-40b1-ba84-855d45b00dc0" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.749168 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rzvjq" event={"ID":"8ab7cf82-a1db-4139-82eb-9d05c0df6a50","Type":"ContainerStarted","Data":"70e0e4e4aae9a9dc5ec32ada1cff6c00c97878ea3d5767eb461fc0e6ccf9b428"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.777541 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" podStartSLOduration=156.777516351 podStartE2EDuration="2m36.777516351s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:31.775936426 +0000 UTC m=+213.096354633" watchObservedRunningTime="2026-03-20 09:00:31.777516351 +0000 UTC m=+213.097934548" Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.779122 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.779813 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.279792953 +0000 UTC m=+213.600211150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.785443 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.787207 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.287189592 +0000 UTC m=+213.607607789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.791880 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" event={"ID":"a9f367b2-d0b3-4a80-933f-68bf11e63791","Type":"ContainerStarted","Data":"0a12e38f2d188aa3ff73040821e951e7b7c1e50600b35e2ba3035095d967c700"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.851399 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" event={"ID":"31262c04-c5c9-4b06-afd8-f005d271819a","Type":"ContainerStarted","Data":"f974ae9158ea42fdd3f3185c85fb21d1dee65aed260efc79af1183a8f60f2d0b"} Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.888060 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.893657 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.393619999 +0000 UTC m=+213.714038196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:31 crc kubenswrapper[4858]: I0320 09:00:31.903459 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:31 crc kubenswrapper[4858]: E0320 09:00:31.907335 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.40729999 +0000 UTC m=+213.727718187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.023807 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.026284 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.526260432 +0000 UTC m=+213.846678629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.068753 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.089094 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-vtwn4" podStartSLOduration=157.089069784 podStartE2EDuration="2m37.089069784s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:32.085096124 +0000 UTC m=+213.405514321" watchObservedRunningTime="2026-03-20 09:00:32.089069784 +0000 UTC m=+213.409487971" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.115881 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-jsp8n" podStartSLOduration=157.115855925 podStartE2EDuration="2m37.115855925s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:32.11298724 +0000 UTC m=+213.433405447" watchObservedRunningTime="2026-03-20 09:00:32.115855925 +0000 UTC m=+213.436274142" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.122613 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wr84h" event={"ID":"91be84d3-8196-44bb-8a88-e9e6548377a1","Type":"ContainerStarted","Data":"273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152"} Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.129684 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.138673 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.638640584 +0000 UTC m=+213.959058781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.144971 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" event={"ID":"3b3bd918-04b9-4371-933a-609e9add5512","Type":"ContainerStarted","Data":"ad9a868a1efa3260860f9a49986f9372c09fdd86fd495868b7021ea13d85f956"} Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.188268 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" event={"ID":"afdaaa04-0b97-4c8c-8699-a620209b9202","Type":"ContainerStarted","Data":"18b4e0ef2c1c5ee27071385f60c9fff2506b4e9c5eb70b668d74a196a556e5e7"} Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.189093 4858 patch_prober.go:28] interesting pod/downloads-7954f5f757-j6mmm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.189126 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-j6mmm" podUID="e07edf68-41a8-4175-adc0-163e46620ab4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.212101 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pbg4m" podStartSLOduration=157.212076698 podStartE2EDuration="2m37.212076698s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:32.164148056 +0000 UTC m=+213.484566253" watchObservedRunningTime="2026-03-20 09:00:32.212076698 +0000 UTC m=+213.532494895" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.243160 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.247091 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.747062237 +0000 UTC m=+214.067480434 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.272237 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" podStartSLOduration=157.27219817 podStartE2EDuration="2m37.27219817s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:32.214669568 +0000 UTC m=+213.535087765" watchObservedRunningTime="2026-03-20 09:00:32.27219817 +0000 UTC m=+213.592616367" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.276929 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bkvnd" podStartSLOduration=157.276919717 podStartE2EDuration="2m37.276919717s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:32.272406274 +0000 UTC m=+213.592824481" watchObservedRunningTime="2026-03-20 09:00:32.276919717 +0000 UTC m=+213.597337934" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.311559 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-wr84h" podStartSLOduration=157.311529576 podStartE2EDuration="2m37.311529576s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:32.310518673 +0000 UTC m=+213.630936890" watchObservedRunningTime="2026-03-20 09:00:32.311529576 +0000 UTC m=+213.631947783" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.344997 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.345491 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.8454737 +0000 UTC m=+214.165891897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.433043 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2"] Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.445494 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.445859 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:32.945831408 +0000 UTC m=+214.266249605 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.463348 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j"] Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.512079 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f"] Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.547346 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.547881 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.047854374 +0000 UTC m=+214.368272751 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.555597 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:32 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:32 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:32 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.555687 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.558900 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22"] Mar 20 09:00:32 crc kubenswrapper[4858]: W0320 09:00:32.574213 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fae0532_0891_4c58_abec_e48437904f40.slice/crio-cf9e452af5e9ef7153d362cef548c5419589878ce4a7abed7c5acfc14509d48c WatchSource:0}: Error finding container cf9e452af5e9ef7153d362cef548c5419589878ce4a7abed7c5acfc14509d48c: Status 404 returned error can't find the container with id cf9e452af5e9ef7153d362cef548c5419589878ce4a7abed7c5acfc14509d48c Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.646467 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj"] Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.649625 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.650099 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.150069644 +0000 UTC m=+214.470487831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.751006 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.751351 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.251336823 +0000 UTC m=+214.571755020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.812273 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs"] Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.861666 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4s5wf"] Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.864830 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.866727 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.366695993 +0000 UTC m=+214.687114190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:32 crc kubenswrapper[4858]: I0320 09:00:32.974836 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:32 crc kubenswrapper[4858]: E0320 09:00:32.975281 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.475262308 +0000 UTC m=+214.795680505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.022710 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-tbsv5"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.034300 4858 ???:1] "http: TLS handshake error from 192.168.126.11:46994: no serving certificate available for the kubelet" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.043419 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-9rv48"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.075358 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.076005 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.575977254 +0000 UTC m=+214.896395441 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.111337 4858 ???:1] "http: TLS handshake error from 192.168.126.11:47010: no serving certificate available for the kubelet" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.119906 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-h64jx"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.179874 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.180263 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.680247831 +0000 UTC m=+215.000666028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.211110 4858 ???:1] "http: TLS handshake error from 192.168.126.11:47022: no serving certificate available for the kubelet" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.251405 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" event={"ID":"2fae0532-0891-4c58-abec-e48437904f40","Type":"ContainerStarted","Data":"cf9e452af5e9ef7153d362cef548c5419589878ce4a7abed7c5acfc14509d48c"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.253222 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-b58fr"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.265933 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.281582 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.281739 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.781717414 +0000 UTC m=+215.102135611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.282157 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.282615 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.782604964 +0000 UTC m=+215.103023161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.293242 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" event={"ID":"a91bc770-3810-4e8e-ac89-4b321be44b3c","Type":"ContainerStarted","Data":"d7e5026af71439758b6135ce0b9c803b134000c1a002935686f88a5220b818b7"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.315061 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" event={"ID":"e2f99651-1c5a-4f42-a46e-af580ec9b4eb","Type":"ContainerStarted","Data":"d8dafec60f04567dd9503e45024f49c338a713f55c3c1fda4929004edd3e7072"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.320219 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" event={"ID":"3c53fc26-4e6d-4d8f-bb46-59987bcc746f","Type":"ContainerStarted","Data":"cbc7f969e29a7f997ab2acceab6acad7b57e9de13839cb206eb37285c6724d19"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.322259 4858 ???:1] "http: TLS handshake error from 192.168.126.11:47026: no serving certificate available for the kubelet" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.329466 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" event={"ID":"effb9468-b572-4eb1-84df-15e7b0201dbf","Type":"ContainerStarted","Data":"abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.330045 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.341781 4858 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-sg9cs container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.342083 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" podUID="effb9468-b572-4eb1-84df-15e7b0201dbf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.341843 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:33 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:33 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:33 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.342616 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.363532 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" event={"ID":"80c13002-5ff3-43ea-be87-e1b2ecf4431a","Type":"ContainerStarted","Data":"6a3014c91f67d42f1c355fa9bbb1be8e37cd2fb24e0ef689026a78bcdbd26be0"} Mar 20 09:00:33 crc kubenswrapper[4858]: W0320 09:00:33.377924 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod978fa44d_6fb3_4775_ad8f_d81a582b521e.slice/crio-ceb55f1ce0a7a879a86a4e6f290755b51d4ba4cbd809ca1e4800d13ed4cde6f5 WatchSource:0}: Error finding container ceb55f1ce0a7a879a86a4e6f290755b51d4ba4cbd809ca1e4800d13ed4cde6f5: Status 404 returned error can't find the container with id ceb55f1ce0a7a879a86a4e6f290755b51d4ba4cbd809ca1e4800d13ed4cde6f5 Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.382745 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.384993 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.884961908 +0000 UTC m=+215.205380105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.385417 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" event={"ID":"59f9f7fc-da90-46ec-b360-1eee512a4416","Type":"ContainerStarted","Data":"fe74280b71e1a07d7092808ccaf52f1d74ab2d50fe68ff22ab570232cbf7273f"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.387693 4858 generic.go:334] "Generic (PLEG): container finished" podID="055f16c2-9ca1-4078-82b9-48aa9a4399ad" containerID="d5a3621adf58a56531f76ad8f038a294efb70fa10c6ed3f5613712c6c848683b" exitCode=0 Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.388321 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" event={"ID":"055f16c2-9ca1-4078-82b9-48aa9a4399ad","Type":"ContainerDied","Data":"d5a3621adf58a56531f76ad8f038a294efb70fa10c6ed3f5613712c6c848683b"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.388431 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-vw55r" podStartSLOduration=158.388416807 podStartE2EDuration="2m38.388416807s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:33.352648761 +0000 UTC m=+214.673066958" watchObservedRunningTime="2026-03-20 09:00:33.388416807 +0000 UTC m=+214.708835004" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.389840 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" podStartSLOduration=158.389834239 podStartE2EDuration="2m38.389834239s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:33.385114712 +0000 UTC m=+214.705532929" watchObservedRunningTime="2026-03-20 09:00:33.389834239 +0000 UTC m=+214.710252436" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.390677 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" event={"ID":"48bc9663-4753-4ad1-b0f4-2414dc389098","Type":"ContainerStarted","Data":"67d274d26cd83cf7bf2890fbefae53e84d1f001fb6281cbb5618c0c8aa9826e3"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.409428 4858 ???:1] "http: TLS handshake error from 192.168.126.11:47042: no serving certificate available for the kubelet" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.410308 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" event={"ID":"960976f4-1bea-423a-b4fc-09b08a60ba0d","Type":"ContainerStarted","Data":"0176cae30480c5a6867caf12e9c6885efbb568ccccf2db7cbf68bd5746eb8cb6"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.414692 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" event={"ID":"99346cc6-9090-4e06-beb0-d64a92bd2813","Type":"ContainerStarted","Data":"acc6855e5c6080ee57cb3a4b4fd2934287a0ba1eda8954d03cbf2af03fa39f35"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.418028 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" event={"ID":"84a3f478-6ec6-4ef6-983e-acdeeb3b475c","Type":"ContainerStarted","Data":"d2b4bf87dc906cb85e330f95fc91254c14043ef384779ba0d1f7aa73255a7e8d"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.420058 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" event={"ID":"b51deb4e-ca50-41d4-8b00-bb996f8e7782","Type":"ContainerStarted","Data":"cc5dcf9fab5fde9bc933aaac378b633c22eb729c28acec4935a1b457f3d37ea8"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.423587 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-zdvzk" podStartSLOduration=158.423565158 podStartE2EDuration="2m38.423565158s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:33.420653302 +0000 UTC m=+214.741071509" watchObservedRunningTime="2026-03-20 09:00:33.423565158 +0000 UTC m=+214.743983355" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.431908 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" event={"ID":"c245d181-9680-448c-a0c6-32f5d54811f7","Type":"ContainerStarted","Data":"17d517f688915fe1802b813fdd3e89cad05a575bcbf0519f8b43fbf4c6728264"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.446545 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" event={"ID":"d8f37a14-0144-4154-a087-126fde1633eb","Type":"ContainerStarted","Data":"fe924a2f8c500c5a40f91694fab8c4b75b9ee5cad87fd09b14576ce594d745c0"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.454033 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.463222 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.466721 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.467009 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" event={"ID":"74fe10ec-a162-4c93-b2d3-1a80745e7fcc","Type":"ContainerStarted","Data":"5c8d754c20df28d8c41a4c9f38670efb315b937f17d8c6c667779683b96d0d23"} Mar 20 09:00:33 crc kubenswrapper[4858]: W0320 09:00:33.472857 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod631923d2_1540_4a96_889a_d6f39d28ef1b.slice/crio-b28a544b72387d753f8d479382321bf4a13e4eb66f17d1d11c1061f55f0969e0 WatchSource:0}: Error finding container b28a544b72387d753f8d479382321bf4a13e4eb66f17d1d11c1061f55f0969e0: Status 404 returned error can't find the container with id b28a544b72387d753f8d479382321bf4a13e4eb66f17d1d11c1061f55f0969e0 Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.483755 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" podStartSLOduration=158.48373317 podStartE2EDuration="2m38.48373317s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:33.48287909 +0000 UTC m=+214.803297287" watchObservedRunningTime="2026-03-20 09:00:33.48373317 +0000 UTC m=+214.804151367" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.485445 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.487063 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:33.987047915 +0000 UTC m=+215.307466302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.491435 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" event={"ID":"563363e9-546c-4754-93b8-c274c58779b0","Type":"ContainerStarted","Data":"57b4fed6bdea14b76e281668909a4bf3e48ad324c93196fe94eb129c9e080084"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.491894 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-c47vz"] Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.518716 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" event={"ID":"6db678cf-767f-4339-90db-09aa1fe57983","Type":"ContainerStarted","Data":"2a5d7b6c95d0c835b32ae190837ff7944dfcf4d124e7023f125ea0f45069aeae"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.536064 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" event={"ID":"69e5a3c1-6c66-42f5-a122-63c4d2838aca","Type":"ContainerStarted","Data":"b573c6ceb3706fb472c3987a1bfc96b91e3dd4c2bdf55414185f517108eaa742"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.537694 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.542044 4858 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-rhqsx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.542090 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" podUID="69e5a3c1-6c66-42f5-a122-63c4d2838aca" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.543975 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" event={"ID":"dc30d82f-f02c-41c4-9c6e-94e663fa8712","Type":"ContainerStarted","Data":"6ffbedaf93188623f394f48d867ad919b5067389cdd459703d2cd41e986e7e38"} Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.562157 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-cwnt6" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.567719 4858 ???:1] "http: TLS handshake error from 192.168.126.11:51066: no serving certificate available for the kubelet" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.587297 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.599089 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.099044098 +0000 UTC m=+215.419462295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: W0320 09:00:33.599253 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-28df59e13a25b953c496d5841eeaa6982adc92cdc699be8c4595b48f52184c8e WatchSource:0}: Error finding container 28df59e13a25b953c496d5841eeaa6982adc92cdc699be8c4595b48f52184c8e: Status 404 returned error can't find the container with id 28df59e13a25b953c496d5841eeaa6982adc92cdc699be8c4595b48f52184c8e Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.624257 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" podStartSLOduration=158.624227012 podStartE2EDuration="2m38.624227012s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:33.579182076 +0000 UTC m=+214.899600283" watchObservedRunningTime="2026-03-20 09:00:33.624227012 +0000 UTC m=+214.944645209" Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.689883 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.189849689 +0000 UTC m=+215.510267886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.689299 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.754207 4858 ???:1] "http: TLS handshake error from 192.168.126.11:51074: no serving certificate available for the kubelet" Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.795534 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.795793 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.295756694 +0000 UTC m=+215.616174891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.795866 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.796368 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.296350497 +0000 UTC m=+215.616768694 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.897855 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.898095 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.398048565 +0000 UTC m=+215.718466802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:33 crc kubenswrapper[4858]: I0320 09:00:33.898254 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:33 crc kubenswrapper[4858]: E0320 09:00:33.898895 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.398879274 +0000 UTC m=+215.719297501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.000247 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.000550 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.500528212 +0000 UTC m=+215.820946419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.102383 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.102998 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.602968077 +0000 UTC m=+215.923386284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.104355 4858 ???:1] "http: TLS handshake error from 192.168.126.11:51082: no serving certificate available for the kubelet" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.204405 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.205667 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.705627118 +0000 UTC m=+216.026045315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.307542 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.308029 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.808006511 +0000 UTC m=+216.128424708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.351736 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:34 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:34 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:34 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.352164 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.414112 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.414610 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:34.914595111 +0000 UTC m=+216.235013308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.515882 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.518180 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.018167082 +0000 UTC m=+216.338585279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.550852 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" event={"ID":"e587c0c9-bf99-4f08-96de-e3a386af8b8a","Type":"ContainerStarted","Data":"0cf4ded73e07d809052ab2a35afe6064c266546600f22836bdc98588cf915814"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.550901 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" event={"ID":"e587c0c9-bf99-4f08-96de-e3a386af8b8a","Type":"ContainerStarted","Data":"b06866fe21c51791e92c47735b80591df25b76a0f770b6a0ba459692fc6f3cd2"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.559967 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" event={"ID":"2cc056e7-0895-499c-acb5-0c82b7a8b900","Type":"ContainerStarted","Data":"a4f306a749de27c70bb36e9bf6929bc9e956f43970fd435d809d25853fdbbea0"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.572251 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"daa5f453f7e2ea59db0bac4522a14b391320783f87afd1dceca8461623049cd5"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.572827 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"08aebc41cc03ea170e5f3b4550ccda393cf15deaa0a3b908987513fd6d0e378c"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.573899 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.585609 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rzvjq" event={"ID":"8ab7cf82-a1db-4139-82eb-9d05c0df6a50","Type":"ContainerStarted","Data":"2db95de85597bbe9947a76f6533c22030461525f25857c43a22e37c0f9d60b24"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.601868 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-9rv48" podStartSLOduration=159.60185023 podStartE2EDuration="2m39.60185023s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.573511094 +0000 UTC m=+215.893929301" watchObservedRunningTime="2026-03-20 09:00:34.60185023 +0000 UTC m=+215.922268427" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.619844 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.621224 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.121205982 +0000 UTC m=+216.441624179 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.675942 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c47vz" event={"ID":"95a6716d-90d6-4984-a85c-eab9192d4d75","Type":"ContainerStarted","Data":"f5030dddb7ff65759af06aa2f7158c8cecc5d8800bd7595a57335df81a44c601"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.685846 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rzvjq" podStartSLOduration=7.685817415 podStartE2EDuration="7.685817415s" podCreationTimestamp="2026-03-20 09:00:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.682305395 +0000 UTC m=+216.002723592" watchObservedRunningTime="2026-03-20 09:00:34.685817415 +0000 UTC m=+216.006235612" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.704384 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" event={"ID":"dc30d82f-f02c-41c4-9c6e-94e663fa8712","Type":"ContainerStarted","Data":"7fef26642390cd18255aa4c42e9a96ea3b120bb2e84b0c720dd61c3cbe9c6739"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.715912 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" event={"ID":"563363e9-546c-4754-93b8-c274c58779b0","Type":"ContainerStarted","Data":"22b07d97cb4b60282b098a31bf86bbe4c339023fb45d0c2bca642c2229074fd1"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.717967 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.720023 4858 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nvr22 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.720080 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" podUID="563363e9-546c-4754-93b8-c274c58779b0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.725798 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.727232 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.227202668 +0000 UTC m=+216.547620865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.730953 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-qw67f" podStartSLOduration=159.730921123 podStartE2EDuration="2m39.730921123s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.727605947 +0000 UTC m=+216.048024154" watchObservedRunningTime="2026-03-20 09:00:34.730921123 +0000 UTC m=+216.051339320" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.758354 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" podStartSLOduration=159.758338448 podStartE2EDuration="2m39.758338448s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.757639202 +0000 UTC m=+216.078057399" watchObservedRunningTime="2026-03-20 09:00:34.758338448 +0000 UTC m=+216.078756645" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.758690 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" event={"ID":"b51deb4e-ca50-41d4-8b00-bb996f8e7782","Type":"ContainerStarted","Data":"433983995abce7645763a69512b7efb829cac5de46f154c32d1284a17a409e3e"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.759714 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" event={"ID":"631923d2-1540-4a96-889a-d6f39d28ef1b","Type":"ContainerStarted","Data":"b28a544b72387d753f8d479382321bf4a13e4eb66f17d1d11c1061f55f0969e0"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.770629 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" event={"ID":"960976f4-1bea-423a-b4fc-09b08a60ba0d","Type":"ContainerStarted","Data":"8f5236662020ece1b168b23dad1c53711c4148f12ebc1e55fd80ae9e049c1f1c"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.774632 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" event={"ID":"48bc9663-4753-4ad1-b0f4-2414dc389098","Type":"ContainerStarted","Data":"30af13296b4ed4b4b44d5fdd3e396fca834991c683b27d0c7eb0bf7308530600"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.792399 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-msrvs" podStartSLOduration=159.792382524 podStartE2EDuration="2m39.792382524s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.790802769 +0000 UTC m=+216.111220956" watchObservedRunningTime="2026-03-20 09:00:34.792382524 +0000 UTC m=+216.112800721" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.794559 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" event={"ID":"6db678cf-767f-4339-90db-09aa1fe57983","Type":"ContainerStarted","Data":"694f4f0bd2916e4642d0bd4c25c58fd13df46f62c42c8d9690cb5202b9a7a500"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.797102 4858 ???:1] "http: TLS handshake error from 192.168.126.11:51096: no serving certificate available for the kubelet" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.815260 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-fzvwc" podStartSLOduration=159.815239216 podStartE2EDuration="2m39.815239216s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.809686649 +0000 UTC m=+216.130104836" watchObservedRunningTime="2026-03-20 09:00:34.815239216 +0000 UTC m=+216.135657413" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.816018 4858 generic.go:334] "Generic (PLEG): container finished" podID="990e152e-f7bc-4811-bc9a-6954a09b166a" containerID="f6718bcddc10a7265dbf5f38a834377b2b5831cd1f4f2209e3b6048c5f72cb5d" exitCode=0 Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.816167 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" event={"ID":"990e152e-f7bc-4811-bc9a-6954a09b166a","Type":"ContainerDied","Data":"f6718bcddc10a7265dbf5f38a834377b2b5831cd1f4f2209e3b6048c5f72cb5d"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.827049 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.829188 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.329168293 +0000 UTC m=+216.649586490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.850512 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bv8pj" podStartSLOduration=159.850472609 podStartE2EDuration="2m39.850472609s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.848131245 +0000 UTC m=+216.168549462" watchObservedRunningTime="2026-03-20 09:00:34.850472609 +0000 UTC m=+216.170890806" Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.864225 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" event={"ID":"2fae0532-0891-4c58-abec-e48437904f40","Type":"ContainerStarted","Data":"8004458ea1b540e09c35b96a2c561bf39779b4d7c42765627f12b0589006df02"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.873625 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" event={"ID":"978fa44d-6fb3-4775-ad8f-d81a582b521e","Type":"ContainerStarted","Data":"ceb55f1ce0a7a879a86a4e6f290755b51d4ba4cbd809ca1e4800d13ed4cde6f5"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.878704 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" event={"ID":"16f8cb13-edc5-403c-bc8c-2fcd585139b0","Type":"ContainerStarted","Data":"96ab58e5dba1c58fdc86bb9cc294e07645e51539cbdbe76567b036153a79f7eb"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.881946 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" event={"ID":"c245d181-9680-448c-a0c6-32f5d54811f7","Type":"ContainerStarted","Data":"7b0a898a935be1f32883087ff057d77a18afd9768a795d9c34ca6781ba32e7e6"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.907345 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" event={"ID":"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0","Type":"ContainerStarted","Data":"2e4781a98de7c244c6a0629966cf1301b7e26f5833cd78f2f99f3f1684a6b9d8"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.907770 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" event={"ID":"d4193d0b-c02e-4fd4-a3a1-0b6a0770fef0","Type":"ContainerStarted","Data":"c2f525ac268ff9dcfdb72b1cd97da4ba85897988318c63b2b1b7d0a50e499f7c"} Mar 20 09:00:34 crc kubenswrapper[4858]: I0320 09:00:34.945151 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:34 crc kubenswrapper[4858]: E0320 09:00:34.946771 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.446751763 +0000 UTC m=+216.767169960 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.035901 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xm7ws" podStartSLOduration=160.035867455 podStartE2EDuration="2m40.035867455s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:34.943298825 +0000 UTC m=+216.263717032" watchObservedRunningTime="2026-03-20 09:00:35.035867455 +0000 UTC m=+216.356285672" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.046166 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.046676 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.546656641 +0000 UTC m=+216.867074838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.077409 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s487j" podStartSLOduration=160.077382982 podStartE2EDuration="2m40.077382982s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.011342157 +0000 UTC m=+216.331760364" watchObservedRunningTime="2026-03-20 09:00:35.077382982 +0000 UTC m=+216.397801179" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.077807 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rndqc" podStartSLOduration=160.077802031 podStartE2EDuration="2m40.077802031s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.076179635 +0000 UTC m=+216.396597842" watchObservedRunningTime="2026-03-20 09:00:35.077802031 +0000 UTC m=+216.398220228" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.090989 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"abdb89e335b68da3b36946e554cb935081ff611ac1230fd54ed3fa4e1e68790d"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.091060 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"24f70eb443d873789b22551efee49b8a2819546d4b5d7d1e25a2fac4c063d287"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.150369 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.150730 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.650718894 +0000 UTC m=+216.971137081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.189338 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"28df59e13a25b953c496d5841eeaa6982adc92cdc699be8c4595b48f52184c8e"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.236613 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" event={"ID":"80c13002-5ff3-43ea-be87-e1b2ecf4431a","Type":"ContainerStarted","Data":"2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.237404 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.244947 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b58fr" event={"ID":"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0","Type":"ContainerStarted","Data":"109893bb8a6a6ee5279d1e81fa9c12ed7ff56e381c099d34a8a54a300cb55b5a"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.245000 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-b58fr" event={"ID":"4c2dfb4c-a6cb-4736-a668-020e93ffe5f0","Type":"ContainerStarted","Data":"8812d641eec05c03786fd7172210f287d82e35dd6456c36efd188dbf9524e2b6"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.250370 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" event={"ID":"6287e483-4f8f-4be6-840c-a42d3420d3a5","Type":"ContainerStarted","Data":"82103e35bbaf6f919bd4e5ab16f0d75cb283ff44468c97ad77995a79607e251f"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.251187 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.253021 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.752996566 +0000 UTC m=+217.073414833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.253501 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.258916 4858 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4s5wf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.258969 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.293989 4858 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vk2kc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.294081 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" podUID="6287e483-4f8f-4be6-840c-a42d3420d3a5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.294272 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" event={"ID":"a91bc770-3810-4e8e-ac89-4b321be44b3c","Type":"ContainerStarted","Data":"1633ad5ed4e93d83d0deb5fa99f1ac3fe79efd362b10d2c294687930d7c86382"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.313923 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" event={"ID":"a9f367b2-d0b3-4a80-933f-68bf11e63791","Type":"ContainerStarted","Data":"740cae4298efab318dad038c9ccff9acbbd97a0a2ed3ed5670209d0c91ac352a"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.314143 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.315559 4858 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2c66r container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.315629 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" podUID="a9f367b2-d0b3-4a80-933f-68bf11e63791" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.316376 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-b58fr" podStartSLOduration=9.316348949 podStartE2EDuration="9.316348949s" podCreationTimestamp="2026-03-20 09:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.31374316 +0000 UTC m=+216.634161367" watchObservedRunningTime="2026-03-20 09:00:35.316348949 +0000 UTC m=+216.636767146" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.340035 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" event={"ID":"afdaaa04-0b97-4c8c-8699-a620209b9202","Type":"ContainerStarted","Data":"da95d55342890bff9a728a242c9f1b6e97fcd08141bd87bc3dc9b2a286790046"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.340459 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:35 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:35 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:35 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.340555 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.356696 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.361975 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.86194659 +0000 UTC m=+217.182364787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.371963 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" event={"ID":"e2f99651-1c5a-4f42-a46e-af580ec9b4eb","Type":"ContainerStarted","Data":"9edad6f6c795d158c3b286c6e7488addf6000860aebe6227f8599eb41ceb7ddf"} Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.395424 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-rhqsx" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.398793 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.418070 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" podStartSLOduration=160.418010288 podStartE2EDuration="2m40.418010288s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.373967423 +0000 UTC m=+216.694385620" watchObservedRunningTime="2026-03-20 09:00:35.418010288 +0000 UTC m=+216.738428485" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.419295 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" podStartSLOduration=160.419287496 podStartE2EDuration="2m40.419287496s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.418686333 +0000 UTC m=+216.739104530" watchObservedRunningTime="2026-03-20 09:00:35.419287496 +0000 UTC m=+216.739705693" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.465400 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.467023 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:35.967002994 +0000 UTC m=+217.287421381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.507473 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-468wl" podStartSLOduration=160.507454147 podStartE2EDuration="2m40.507454147s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.505221506 +0000 UTC m=+216.825639703" watchObservedRunningTime="2026-03-20 09:00:35.507454147 +0000 UTC m=+216.827872344" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.508049 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" podStartSLOduration=160.50804326 podStartE2EDuration="2m40.50804326s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.455453101 +0000 UTC m=+216.775871308" watchObservedRunningTime="2026-03-20 09:00:35.50804326 +0000 UTC m=+216.828461457" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.544997 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" podStartSLOduration=160.544971562 podStartE2EDuration="2m40.544971562s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.54397747 +0000 UTC m=+216.864395667" watchObservedRunningTime="2026-03-20 09:00:35.544971562 +0000 UTC m=+216.865389759" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.572493 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.572928 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.072911689 +0000 UTC m=+217.393329886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.587140 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" podStartSLOduration=160.587124863 podStartE2EDuration="2m40.587124863s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.584762179 +0000 UTC m=+216.905180386" watchObservedRunningTime="2026-03-20 09:00:35.587124863 +0000 UTC m=+216.907543060" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.635735 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-dcfwn" podStartSLOduration=160.63571163 podStartE2EDuration="2m40.63571163s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.635293751 +0000 UTC m=+216.955711948" watchObservedRunningTime="2026-03-20 09:00:35.63571163 +0000 UTC m=+216.956129827" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.676939 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.677334 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.177296969 +0000 UTC m=+217.497715166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.741510 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" podStartSLOduration=35.741488502 podStartE2EDuration="35.741488502s" podCreationTimestamp="2026-03-20 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:35.702570695 +0000 UTC m=+217.022988892" watchObservedRunningTime="2026-03-20 09:00:35.741488502 +0000 UTC m=+217.061906699" Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.780788 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.781110 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.281098535 +0000 UTC m=+217.601516732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.882215 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.882523 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.382473246 +0000 UTC m=+217.702891443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.882966 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.883575 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.383565951 +0000 UTC m=+217.703984148 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.984305 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.984523 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.484491742 +0000 UTC m=+217.804909939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:35 crc kubenswrapper[4858]: I0320 09:00:35.984887 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:35 crc kubenswrapper[4858]: E0320 09:00:35.985269 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.485252889 +0000 UTC m=+217.805671076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.087243 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.087551 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.587487351 +0000 UTC m=+217.907905548 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.088031 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.088580 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.588557575 +0000 UTC m=+217.908975772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.180687 4858 ???:1] "http: TLS handshake error from 192.168.126.11:51108: no serving certificate available for the kubelet" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.189847 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.190417 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.690389496 +0000 UTC m=+218.010807693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.292264 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.292821 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.792800711 +0000 UTC m=+218.113218908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.340490 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:36 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:36 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:36 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.340570 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.391097 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" event={"ID":"2cc056e7-0895-499c-acb5-0c82b7a8b900","Type":"ContainerStarted","Data":"9597358e04be2e55a592a73da544b4b6a6ad0ed1b5a45e1844b67bc03bd51a7d"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.391182 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" event={"ID":"2cc056e7-0895-499c-acb5-0c82b7a8b900","Type":"ContainerStarted","Data":"24519a83362699808c7a3f7964bc24388c3f96b3f8a30f75bc33bf4fd3231ccf"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.392855 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.393336 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.893288442 +0000 UTC m=+218.213706649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.399273 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" event={"ID":"6db678cf-767f-4339-90db-09aa1fe57983","Type":"ContainerStarted","Data":"7ea78460811245c1b740522a635b907a20286fe2c69e3e94eff27152570eac06"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.399503 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.408016 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" event={"ID":"16f8cb13-edc5-403c-bc8c-2fcd585139b0","Type":"ContainerStarted","Data":"921a96a2d773b447588f0fd2bfd447fbd1be143c64b59b95fd0a3566867dd4b4"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.408134 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" event={"ID":"16f8cb13-edc5-403c-bc8c-2fcd585139b0","Type":"ContainerStarted","Data":"d625498e2c4a51b03f8376b0da4cb1322f0235cba48547a97f5b6b18c86a7416"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.414652 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" event={"ID":"6287e483-4f8f-4be6-840c-a42d3420d3a5","Type":"ContainerStarted","Data":"d4cfa1880ecbacf0514f8ca156be31ad4a7a280b0eddbd5119c2c7132b0ab92a"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.415648 4858 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-vk2kc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.415744 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" podUID="6287e483-4f8f-4be6-840c-a42d3420d3a5" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.420400 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" event={"ID":"631923d2-1540-4a96-889a-d6f39d28ef1b","Type":"ContainerStarted","Data":"5bca4b90196d60450b1a21f78c4a3a8aa7711b211b0e9d9c4a81aa268b9bff5b"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.430902 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" event={"ID":"055f16c2-9ca1-4078-82b9-48aa9a4399ad","Type":"ContainerStarted","Data":"166a7fe442642bb624b1bcd59dd8da5f4c306ba7a6b8b637c4804ceb57f3d6d6"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.462432 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-z9dfb" event={"ID":"afdaaa04-0b97-4c8c-8699-a620209b9202","Type":"ContainerStarted","Data":"4a4960d7e93af7b367e2b4b3aa2f0ed9c7724957409535d3a1b10620fffe725f"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.471131 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6mj9n" podStartSLOduration=161.471113937 podStartE2EDuration="2m41.471113937s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:36.469373626 +0000 UTC m=+217.789791854" watchObservedRunningTime="2026-03-20 09:00:36.471113937 +0000 UTC m=+217.791532134" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.471245 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-h64jx" podStartSLOduration=161.471241989 podStartE2EDuration="2m41.471241989s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:36.42391724 +0000 UTC m=+217.744335437" watchObservedRunningTime="2026-03-20 09:00:36.471241989 +0000 UTC m=+217.791660186" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.483107 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" event={"ID":"990e152e-f7bc-4811-bc9a-6954a09b166a","Type":"ContainerStarted","Data":"9df22ae9c77b0e834f62f834dde98e0feb4d14075b09fee87df2ace6cec9632a"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.483864 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.494546 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.496871 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:36.996850663 +0000 UTC m=+218.317269050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.514912 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1d03e4ae4dd452ec045a932e2cda5f51f073f20d05c422911c56805cfabbf4fb"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.542552 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c47vz" event={"ID":"95a6716d-90d6-4984-a85c-eab9192d4d75","Type":"ContainerStarted","Data":"a4cbf39b7dd082718129b52ee95ed048af3fe188d2d6d9cf7994764f89513d2a"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.542602 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-c47vz" event={"ID":"95a6716d-90d6-4984-a85c-eab9192d4d75","Type":"ContainerStarted","Data":"f25d76742b1b418f7dd67e0f45a64daefbad7e418e1c831d33099fd2dc0f081a"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.542649 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.559039 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-rh86k" podStartSLOduration=161.55902163 podStartE2EDuration="2m41.55902163s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:36.557341262 +0000 UTC m=+217.877759459" watchObservedRunningTime="2026-03-20 09:00:36.55902163 +0000 UTC m=+217.879439827" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.559491 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" podStartSLOduration=161.559485961 podStartE2EDuration="2m41.559485961s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:36.510915834 +0000 UTC m=+217.831334031" watchObservedRunningTime="2026-03-20 09:00:36.559485961 +0000 UTC m=+217.879904158" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.563616 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-z9scw" event={"ID":"a91bc770-3810-4e8e-ac89-4b321be44b3c","Type":"ContainerStarted","Data":"07459fcbade5336f837b80fa9f3c9635837b5776c1d230a624c3e6bcedd2ebc1"} Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.565615 4858 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-nvr22 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.565688 4858 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4s5wf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.566607 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.565683 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" podUID="563363e9-546c-4754-93b8-c274c58779b0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.584273 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" podStartSLOduration=161.584249645 podStartE2EDuration="2m41.584249645s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:36.582137277 +0000 UTC m=+217.902555474" watchObservedRunningTime="2026-03-20 09:00:36.584249645 +0000 UTC m=+217.904667842" Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.602737 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.102709967 +0000 UTC m=+218.423128164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.602773 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.603285 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.604303 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hsp4b"] Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.604644 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" podUID="039cac36-f4ed-4282-aa07-ee40ad00df93" containerName="controller-manager" containerID="cri-o://6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259" gracePeriod=30 Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.621750 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.12173068 +0000 UTC m=+218.442148877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.630298 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-c47vz" podStartSLOduration=9.630266184 podStartE2EDuration="9.630266184s" podCreationTimestamp="2026-03-20 09:00:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:36.621492124 +0000 UTC m=+217.941910331" watchObservedRunningTime="2026-03-20 09:00:36.630266184 +0000 UTC m=+217.950684381" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.702420 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" podStartSLOduration=161.702391839 podStartE2EDuration="2m41.702391839s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:36.699178246 +0000 UTC m=+218.019596463" watchObservedRunningTime="2026-03-20 09:00:36.702391839 +0000 UTC m=+218.022810036" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.708626 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.708821 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.208775934 +0000 UTC m=+218.529194131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.709081 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.709614 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.209594623 +0000 UTC m=+218.530012820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.747473 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs"] Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.812991 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.813532 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.313495232 +0000 UTC m=+218.633913439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.813885 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.814330 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.31429695 +0000 UTC m=+218.634715147 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.915486 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.915797 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.415749023 +0000 UTC m=+218.736167220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.915910 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:36 crc kubenswrapper[4858]: E0320 09:00:36.916891 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.416858409 +0000 UTC m=+218.737276606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.944381 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2dv2r"] Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.945567 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.949310 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 09:00:36 crc kubenswrapper[4858]: I0320 09:00:36.969675 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dv2r"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.017268 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.017527 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-catalog-content\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.017622 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqfcc\" (UniqueName: \"kubernetes.io/projected/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-kube-api-access-jqfcc\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.017651 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-utilities\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.017873 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.51781169 +0000 UTC m=+218.838229887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.121489 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.121577 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqfcc\" (UniqueName: \"kubernetes.io/projected/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-kube-api-access-jqfcc\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.121611 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-utilities\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.121637 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-catalog-content\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.121722 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6kdlf"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.122195 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-catalog-content\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.122561 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.622543018 +0000 UTC m=+218.942961205 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.132862 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-utilities\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.133692 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.140897 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.169824 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqfcc\" (UniqueName: \"kubernetes.io/projected/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-kube-api-access-jqfcc\") pod \"certified-operators-2dv2r\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.199393 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6kdlf"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.222634 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.223231 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.723205263 +0000 UTC m=+219.043623470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.270789 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.310674 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.324186 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fll4j\" (UniqueName: \"kubernetes.io/projected/9daba85d-2681-4f74-8094-9db79d723cee-kube-api-access-fll4j\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.324614 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.324772 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-catalog-content\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.324870 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-utilities\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.325505 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.825486605 +0000 UTC m=+219.145904802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.332542 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.337388 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-crt7w"] Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.337739 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="039cac36-f4ed-4282-aa07-ee40ad00df93" containerName="controller-manager" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.337764 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="039cac36-f4ed-4282-aa07-ee40ad00df93" containerName="controller-manager" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.337915 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="039cac36-f4ed-4282-aa07-ee40ad00df93" containerName="controller-manager" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.340424 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.345429 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:37 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:37 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:37 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.345518 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.362353 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crt7w"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.426238 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.426904 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.926856406 +0000 UTC m=+219.247274723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.427598 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-catalog-content\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.427708 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-utilities\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.427782 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fll4j\" (UniqueName: \"kubernetes.io/projected/9daba85d-2681-4f74-8094-9db79d723cee-kube-api-access-fll4j\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.427832 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.428245 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:37.928234957 +0000 UTC m=+219.248653154 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.428908 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-catalog-content\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.429160 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-utilities\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.462550 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fll4j\" (UniqueName: \"kubernetes.io/projected/9daba85d-2681-4f74-8094-9db79d723cee-kube-api-access-fll4j\") pod \"community-operators-6kdlf\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.526520 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4dwft"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.528388 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.529797 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039cac36-f4ed-4282-aa07-ee40ad00df93-serving-cert\") pod \"039cac36-f4ed-4282-aa07-ee40ad00df93\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530304 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530363 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-client-ca\") pod \"039cac36-f4ed-4282-aa07-ee40ad00df93\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530383 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-config\") pod \"039cac36-f4ed-4282-aa07-ee40ad00df93\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530433 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj6td\" (UniqueName: \"kubernetes.io/projected/039cac36-f4ed-4282-aa07-ee40ad00df93-kube-api-access-rj6td\") pod \"039cac36-f4ed-4282-aa07-ee40ad00df93\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530460 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-proxy-ca-bundles\") pod \"039cac36-f4ed-4282-aa07-ee40ad00df93\" (UID: \"039cac36-f4ed-4282-aa07-ee40ad00df93\") " Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530571 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g575q\" (UniqueName: \"kubernetes.io/projected/1b56366e-866a-4139-9b65-3228c5f92d4a-kube-api-access-g575q\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530612 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-catalog-content\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530639 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-catalog-content\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530669 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9gqc\" (UniqueName: \"kubernetes.io/projected/c8456e28-cc53-4820-8bbf-44e27de1dc9b-kube-api-access-n9gqc\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530741 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-utilities\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.530769 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-utilities\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.530903 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.030876657 +0000 UTC m=+219.351294854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.532038 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "039cac36-f4ed-4282-aa07-ee40ad00df93" (UID: "039cac36-f4ed-4282-aa07-ee40ad00df93"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.532085 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-config" (OuterVolumeSpecName: "config") pod "039cac36-f4ed-4282-aa07-ee40ad00df93" (UID: "039cac36-f4ed-4282-aa07-ee40ad00df93"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.532130 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-client-ca" (OuterVolumeSpecName: "client-ca") pod "039cac36-f4ed-4282-aa07-ee40ad00df93" (UID: "039cac36-f4ed-4282-aa07-ee40ad00df93"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.551821 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/039cac36-f4ed-4282-aa07-ee40ad00df93-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "039cac36-f4ed-4282-aa07-ee40ad00df93" (UID: "039cac36-f4ed-4282-aa07-ee40ad00df93"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.552243 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.553463 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/039cac36-f4ed-4282-aa07-ee40ad00df93-kube-api-access-rj6td" (OuterVolumeSpecName: "kube-api-access-rj6td") pod "039cac36-f4ed-4282-aa07-ee40ad00df93" (UID: "039cac36-f4ed-4282-aa07-ee40ad00df93"). InnerVolumeSpecName "kube-api-access-rj6td". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.568691 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dwft"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.606198 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" event={"ID":"978fa44d-6fb3-4775-ad8f-d81a582b521e","Type":"ContainerStarted","Data":"29f9f1f71df378d79791725d902a4da6db057f8cf3f326e2319590590c7b10cb"} Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633263 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g575q\" (UniqueName: \"kubernetes.io/projected/1b56366e-866a-4139-9b65-3228c5f92d4a-kube-api-access-g575q\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633405 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-catalog-content\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633446 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-catalog-content\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633486 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9gqc\" (UniqueName: \"kubernetes.io/projected/c8456e28-cc53-4820-8bbf-44e27de1dc9b-kube-api-access-n9gqc\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633519 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633590 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-utilities\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633623 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-utilities\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633672 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj6td\" (UniqueName: \"kubernetes.io/projected/039cac36-f4ed-4282-aa07-ee40ad00df93-kube-api-access-rj6td\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633687 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633703 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039cac36-f4ed-4282-aa07-ee40ad00df93-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633717 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.633728 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/039cac36-f4ed-4282-aa07-ee40ad00df93-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.636660 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-utilities\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.637046 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-catalog-content\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.637424 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-catalog-content\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.637542 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.137519058 +0000 UTC m=+219.457937425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.638277 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-utilities\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.661711 4858 generic.go:334] "Generic (PLEG): container finished" podID="039cac36-f4ed-4282-aa07-ee40ad00df93" containerID="6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259" exitCode=0 Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.662750 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.668493 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" event={"ID":"039cac36-f4ed-4282-aa07-ee40ad00df93","Type":"ContainerDied","Data":"6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259"} Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.668565 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hsp4b" event={"ID":"039cac36-f4ed-4282-aa07-ee40ad00df93","Type":"ContainerDied","Data":"1aac04c6846e72c5160319abc1239d2160de18d4efacf3b517ec441ee0b0f1af"} Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.668594 4858 scope.go:117] "RemoveContainer" containerID="6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.682364 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9gqc\" (UniqueName: \"kubernetes.io/projected/c8456e28-cc53-4820-8bbf-44e27de1dc9b-kube-api-access-n9gqc\") pod \"certified-operators-crt7w\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.696040 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-vk2kc" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.703660 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g575q\" (UniqueName: \"kubernetes.io/projected/1b56366e-866a-4139-9b65-3228c5f92d4a-kube-api-access-g575q\") pod \"community-operators-4dwft\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.735387 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.735806 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.235783818 +0000 UTC m=+219.556202015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.794665 4858 scope.go:117] "RemoveContainer" containerID="6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.795406 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259\": container with ID starting with 6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259 not found: ID does not exist" containerID="6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.795467 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259"} err="failed to get container status \"6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259\": rpc error: code = NotFound desc = could not find container \"6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259\": container with ID starting with 6319082d5e472eeffdd232ef87de3902479632329d6952cb930f5f2137351259 not found: ID does not exist" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.834389 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hsp4b"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.834475 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hsp4b"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.841949 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.844739 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.344714392 +0000 UTC m=+219.665132589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.851944 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-658bbb8c56-mkjhs"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.852808 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.868397 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.868735 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.870653 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.870800 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.870948 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.871065 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.871900 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.882013 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.888160 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658bbb8c56-mkjhs"] Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.946911 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:37 crc kubenswrapper[4858]: E0320 09:00:37.947476 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.447458795 +0000 UTC m=+219.767876992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:37 crc kubenswrapper[4858]: I0320 09:00:37.978133 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.059411 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-proxy-ca-bundles\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.059457 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa280af1-a47c-4359-897c-08cc65196a53-serving-cert\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.059507 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-977pd\" (UniqueName: \"kubernetes.io/projected/fa280af1-a47c-4359-897c-08cc65196a53-kube-api-access-977pd\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.059532 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-client-ca\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.059561 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.059594 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-config\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.086013 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.585977542 +0000 UTC m=+219.906395739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.162528 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.162762 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-977pd\" (UniqueName: \"kubernetes.io/projected/fa280af1-a47c-4359-897c-08cc65196a53-kube-api-access-977pd\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.162808 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-client-ca\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.162862 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-config\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.162891 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-proxy-ca-bundles\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.162916 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa280af1-a47c-4359-897c-08cc65196a53-serving-cert\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.163352 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="039cac36-f4ed-4282-aa07-ee40ad00df93" path="/var/lib/kubelet/pods/039cac36-f4ed-4282-aa07-ee40ad00df93/volumes" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.165700 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-client-ca\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.166079 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.666058118 +0000 UTC m=+219.986476315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.168353 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-proxy-ca-bundles\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.170038 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-config\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.177727 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa280af1-a47c-4359-897c-08cc65196a53-serving-cert\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.193047 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.195540 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.211955 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-977pd\" (UniqueName: \"kubernetes.io/projected/fa280af1-a47c-4359-897c-08cc65196a53-kube-api-access-977pd\") pod \"controller-manager-658bbb8c56-mkjhs\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.217725 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dv2r"] Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.225361 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.251589 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-nvr22" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.272934 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.274363 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.774351566 +0000 UTC m=+220.094769753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: W0320 09:00:38.291333 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode78c3dad_ee9d_4901_8c08_2db4bd2070cd.slice/crio-4e0d383385daed9a156bdfec4b18c063b022dd4c235f527272b8dcf6e20da973 WatchSource:0}: Error finding container 4e0d383385daed9a156bdfec4b18c063b022dd4c235f527272b8dcf6e20da973: Status 404 returned error can't find the container with id 4e0d383385daed9a156bdfec4b18c063b022dd4c235f527272b8dcf6e20da973 Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.347079 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:38 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:38 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:38 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.347145 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.349158 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6kdlf"] Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.360656 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.373679 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.373767 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.873750023 +0000 UTC m=+220.194168230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.375433 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.376283 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.87625853 +0000 UTC m=+220.196676727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.497889 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.498297 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.998267401 +0000 UTC m=+220.318685598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.498154 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.498634 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.498974 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:38.998960527 +0000 UTC m=+220.319378724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.599830 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.600156 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.100133274 +0000 UTC m=+220.420551471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.602236 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-xbjjz" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.698553 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kdlf" event={"ID":"9daba85d-2681-4f74-8094-9db79d723cee","Type":"ContainerStarted","Data":"ed4dd9dd394d72e705f8cc80dbbe1ac5b44cdde6c7c14acc4d111267aeede9d0"} Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.706071 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.706473 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.206457598 +0000 UTC m=+220.526875795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.719519 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dv2r" event={"ID":"e78c3dad-ee9d-4901-8c08-2db4bd2070cd","Type":"ContainerStarted","Data":"0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca"} Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.719579 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dv2r" event={"ID":"e78c3dad-ee9d-4901-8c08-2db4bd2070cd","Type":"ContainerStarted","Data":"4e0d383385daed9a156bdfec4b18c063b022dd4c235f527272b8dcf6e20da973"} Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.782625 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4dwft"] Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.823470 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" event={"ID":"978fa44d-6fb3-4775-ad8f-d81a582b521e","Type":"ContainerStarted","Data":"b836690508c9705c5610f065f39e3fe9dc54448c37085f172326080cc65830a3"} Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.825530 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.826968 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.326935925 +0000 UTC m=+220.647354122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:38 crc kubenswrapper[4858]: W0320 09:00:38.868909 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b56366e_866a_4139_9b65_3228c5f92d4a.slice/crio-56189411e3d4ebd6d6333111a6e34c5f42a4d5c36779effda3b59e8fdf1380a5 WatchSource:0}: Error finding container 56189411e3d4ebd6d6333111a6e34c5f42a4d5c36779effda3b59e8fdf1380a5: Status 404 returned error can't find the container with id 56189411e3d4ebd6d6333111a6e34c5f42a4d5c36779effda3b59e8fdf1380a5 Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.871072 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" podUID="effb9468-b572-4eb1-84df-15e7b0201dbf" containerName="route-controller-manager" containerID="cri-o://abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6" gracePeriod=30 Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.901487 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-c5kzc" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.907566 4858 ???:1] "http: TLS handshake error from 192.168.126.11:51110: no serving certificate available for the kubelet" Mar 20 09:00:38 crc kubenswrapper[4858]: I0320 09:00:38.927658 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:38 crc kubenswrapper[4858]: E0320 09:00:38.932879 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.432858869 +0000 UTC m=+220.753277246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.038613 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crt7w"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.051243 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.051690 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.551673088 +0000 UTC m=+220.872091285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: W0320 09:00:39.142666 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8456e28_cc53_4820_8bbf_44e27de1dc9b.slice/crio-45d9e21713de5685e5de8d219f56ded96a618ff8782d0482c8a18da0de683f6c WatchSource:0}: Error finding container 45d9e21713de5685e5de8d219f56ded96a618ff8782d0482c8a18da0de683f6c: Status 404 returned error can't find the container with id 45d9e21713de5685e5de8d219f56ded96a618ff8782d0482c8a18da0de683f6c Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.142764 4858 patch_prober.go:28] interesting pod/downloads-7954f5f757-j6mmm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.142831 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-j6mmm" podUID="e07edf68-41a8-4175-adc0-163e46620ab4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.153204 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.153611 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.653599192 +0000 UTC m=+220.974017389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.157950 4858 patch_prober.go:28] interesting pod/downloads-7954f5f757-j6mmm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.158044 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-j6mmm" podUID="e07edf68-41a8-4175-adc0-163e46620ab4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.254068 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.254796 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.754768908 +0000 UTC m=+221.075187105 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.269797 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pbnzw"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.272787 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.277058 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.283958 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.283989 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.285584 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbnzw"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.324734 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.334825 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.364558 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.364626 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pggl\" (UniqueName: \"kubernetes.io/projected/508d2d5b-0a75-4130-a396-9253b685e2cd-kube-api-access-6pggl\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.364675 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-utilities\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.364761 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-catalog-content\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.365753 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.865733978 +0000 UTC m=+221.186152175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.372184 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:39 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:39 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:39 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.372285 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.434575 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658bbb8c56-mkjhs"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.454516 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.454866 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.469010 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.469588 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pggl\" (UniqueName: \"kubernetes.io/projected/508d2d5b-0a75-4130-a396-9253b685e2cd-kube-api-access-6pggl\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.469731 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-utilities\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.469852 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-catalog-content\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.470415 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-catalog-content\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.470581 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:39.970556547 +0000 UTC m=+221.290974744 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.470858 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-utilities\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.475409 4858 patch_prober.go:28] interesting pod/console-f9d7485db-wr84h container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.475473 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wr84h" podUID="91be84d3-8196-44bb-8a88-e9e6548377a1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.490033 4858 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-sg9cs container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.490104 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" podUID="effb9468-b572-4eb1-84df-15e7b0201dbf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.512952 4858 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.519955 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pggl\" (UniqueName: \"kubernetes.io/projected/508d2d5b-0a75-4130-a396-9253b685e2cd-kube-api-access-6pggl\") pod \"redhat-marketplace-pbnzw\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.538572 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mw84k"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.545548 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.571997 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.574099 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.074086028 +0000 UTC m=+221.394504225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.575804 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mw84k"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.597287 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.673432 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.673693 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.173645918 +0000 UTC m=+221.494064115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.673762 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-utilities\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.673937 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.673981 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4jf2\" (UniqueName: \"kubernetes.io/projected/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-kube-api-access-p4jf2\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.674108 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-catalog-content\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.674404 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.174381724 +0000 UTC m=+221.494799911 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.774043 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.775627 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.775864 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.775949 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4jf2\" (UniqueName: \"kubernetes.io/projected/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-kube-api-access-p4jf2\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.776036 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-catalog-content\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.776090 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-utilities\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.776157 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.276119354 +0000 UTC m=+221.596537551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.776832 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-utilities\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.777146 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-catalog-content\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.778968 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.780136 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.780681 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.811239 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4jf2\" (UniqueName: \"kubernetes.io/projected/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-kube-api-access-p4jf2\") pod \"redhat-marketplace-mw84k\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.878544 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.878664 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.878727 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.879298 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.379277576 +0000 UTC m=+221.699695773 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.887659 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.893562 4858 generic.go:334] "Generic (PLEG): container finished" podID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerID="0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca" exitCode=0 Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.893654 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dv2r" event={"ID":"e78c3dad-ee9d-4901-8c08-2db4bd2070cd","Type":"ContainerDied","Data":"0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.896391 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" event={"ID":"fa280af1-a47c-4359-897c-08cc65196a53","Type":"ContainerStarted","Data":"8629c4be903257eea5eda9353696b1fb2c21fba156a98bc6cadb1fd3086c6dab"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.898905 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" event={"ID":"978fa44d-6fb3-4775-ad8f-d81a582b521e","Type":"ContainerStarted","Data":"eace2459410b8df940d1ad5f990848e73ed43885e87eeb57de1d43f08dfac73a"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.917260 4858 generic.go:334] "Generic (PLEG): container finished" podID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerID="f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3" exitCode=0 Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.917366 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crt7w" event={"ID":"c8456e28-cc53-4820-8bbf-44e27de1dc9b","Type":"ContainerDied","Data":"f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.917443 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crt7w" event={"ID":"c8456e28-cc53-4820-8bbf-44e27de1dc9b","Type":"ContainerStarted","Data":"45d9e21713de5685e5de8d219f56ded96a618ff8782d0482c8a18da0de683f6c"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.925265 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g"] Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.925550 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="effb9468-b572-4eb1-84df-15e7b0201dbf" containerName="route-controller-manager" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.925572 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="effb9468-b572-4eb1-84df-15e7b0201dbf" containerName="route-controller-manager" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.925687 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="effb9468-b572-4eb1-84df-15e7b0201dbf" containerName="route-controller-manager" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.926064 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.926453 4858 generic.go:334] "Generic (PLEG): container finished" podID="effb9468-b572-4eb1-84df-15e7b0201dbf" containerID="abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6" exitCode=0 Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.926486 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" event={"ID":"effb9468-b572-4eb1-84df-15e7b0201dbf","Type":"ContainerDied","Data":"abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.926515 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" event={"ID":"effb9468-b572-4eb1-84df-15e7b0201dbf","Type":"ContainerDied","Data":"81daaedce65b3bc7d29a1d604b0acf510de0aebdcf9269fd1d4f8e329dad717c"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.926530 4858 scope.go:117] "RemoveContainer" containerID="abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.926475 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.938155 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.956841 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g"] Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.969082 4858 generic.go:334] "Generic (PLEG): container finished" podID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerID="0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b" exitCode=0 Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.969551 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwft" event={"ID":"1b56366e-866a-4139-9b65-3228c5f92d4a","Type":"ContainerDied","Data":"0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.969575 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwft" event={"ID":"1b56366e-866a-4139-9b65-3228c5f92d4a","Type":"ContainerStarted","Data":"56189411e3d4ebd6d6333111a6e34c5f42a4d5c36779effda3b59e8fdf1380a5"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.974140 4858 generic.go:334] "Generic (PLEG): container finished" podID="9daba85d-2681-4f74-8094-9db79d723cee" containerID="d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd" exitCode=0 Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.975560 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kdlf" event={"ID":"9daba85d-2681-4f74-8094-9db79d723cee","Type":"ContainerDied","Data":"d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd"} Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.983945 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-vfg5z" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.984007 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/effb9468-b572-4eb1-84df-15e7b0201dbf-serving-cert\") pod \"effb9468-b572-4eb1-84df-15e7b0201dbf\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.984095 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tszd\" (UniqueName: \"kubernetes.io/projected/effb9468-b572-4eb1-84df-15e7b0201dbf-kube-api-access-7tszd\") pod \"effb9468-b572-4eb1-84df-15e7b0201dbf\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.984159 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-client-ca\") pod \"effb9468-b572-4eb1-84df-15e7b0201dbf\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.984201 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-config\") pod \"effb9468-b572-4eb1-84df-15e7b0201dbf\" (UID: \"effb9468-b572-4eb1-84df-15e7b0201dbf\") " Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.984302 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.985048 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.985090 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-config\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.985119 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-client-ca\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.985139 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad905103-b4e5-408e-bee8-15591d245f7b-serving-cert\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.985170 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.985251 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hhmb\" (UniqueName: \"kubernetes.io/projected/ad905103-b4e5-408e-bee8-15591d245f7b-kube-api-access-5hhmb\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:39 crc kubenswrapper[4858]: E0320 09:00:39.986469 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.486436998 +0000 UTC m=+221.806855195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.986549 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.986946 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-client-ca" (OuterVolumeSpecName: "client-ca") pod "effb9468-b572-4eb1-84df-15e7b0201dbf" (UID: "effb9468-b572-4eb1-84df-15e7b0201dbf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.987175 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-config" (OuterVolumeSpecName: "config") pod "effb9468-b572-4eb1-84df-15e7b0201dbf" (UID: "effb9468-b572-4eb1-84df-15e7b0201dbf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.992511 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/effb9468-b572-4eb1-84df-15e7b0201dbf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "effb9468-b572-4eb1-84df-15e7b0201dbf" (UID: "effb9468-b572-4eb1-84df-15e7b0201dbf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:00:39 crc kubenswrapper[4858]: I0320 09:00:39.998435 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/effb9468-b572-4eb1-84df-15e7b0201dbf-kube-api-access-7tszd" (OuterVolumeSpecName: "kube-api-access-7tszd") pod "effb9468-b572-4eb1-84df-15e7b0201dbf" (UID: "effb9468-b572-4eb1-84df-15e7b0201dbf"). InnerVolumeSpecName "kube-api-access-7tszd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.017734 4858 scope.go:117] "RemoveContainer" containerID="abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.017767 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:40 crc kubenswrapper[4858]: E0320 09:00:40.018194 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6\": container with ID starting with abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6 not found: ID does not exist" containerID="abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.018218 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6"} err="failed to get container status \"abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6\": rpc error: code = NotFound desc = could not find container \"abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6\": container with ID starting with abf64b39c6e54b36b8ae450a54aa51a96d8fa333a1b50859710e87fb4283cab6 not found: ID does not exist" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.092684 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.092793 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hhmb\" (UniqueName: \"kubernetes.io/projected/ad905103-b4e5-408e-bee8-15591d245f7b-kube-api-access-5hhmb\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.092874 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-config\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.092917 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-client-ca\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.092936 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad905103-b4e5-408e-bee8-15591d245f7b-serving-cert\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.103005 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/effb9468-b572-4eb1-84df-15e7b0201dbf-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.103065 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tszd\" (UniqueName: \"kubernetes.io/projected/effb9468-b572-4eb1-84df-15e7b0201dbf-kube-api-access-7tszd\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.103081 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.103092 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/effb9468-b572-4eb1-84df-15e7b0201dbf-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.110791 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-client-ca\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.111383 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-config\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: E0320 09:00:40.111584 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.611560331 +0000 UTC m=+221.931978528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.116718 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad905103-b4e5-408e-bee8-15591d245f7b-serving-cert\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.154746 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-swvjn"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.183615 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.186258 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hhmb\" (UniqueName: \"kubernetes.io/projected/ad905103-b4e5-408e-bee8-15591d245f7b-kube-api-access-5hhmb\") pod \"route-controller-manager-78cbcb7cb6-88g8g\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.205083 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:40 crc kubenswrapper[4858]: E0320 09:00:40.205652 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.705609506 +0000 UTC m=+222.026027703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.206360 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:40 crc kubenswrapper[4858]: E0320 09:00:40.207256 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.707210662 +0000 UTC m=+222.027628859 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.212688 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.225377 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.243915 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-swvjn"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.261585 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbnzw"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.267481 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.279223 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.317183 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.317505 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-utilities\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.317535 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jczn6\" (UniqueName: \"kubernetes.io/projected/f98a0de8-b0a6-4c33-83b9-831c88485e50-kube-api-access-jczn6\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: E0320 09:00:40.317673 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.81764296 +0000 UTC m=+222.138061157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.317757 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.317779 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-catalog-content\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: E0320 09:00:40.318187 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-20 09:00:40.818180372 +0000 UTC m=+222.138598569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8crqq" (UID: "c2e4d497-a390-4102-961e-8334641b8867") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.330832 4858 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-20T09:00:39.512977274Z","Handler":null,"Name":""} Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.360263 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:40 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:40 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:40 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.360340 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.400677 4858 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.401095 4858 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.418610 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.419021 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-catalog-content\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.419076 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-utilities\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.419100 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jczn6\" (UniqueName: \"kubernetes.io/projected/f98a0de8-b0a6-4c33-83b9-831c88485e50-kube-api-access-jczn6\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.425013 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-catalog-content\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.425251 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-utilities\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.438032 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.438152 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.442621 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-sg9cs"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.477697 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jczn6\" (UniqueName: \"kubernetes.io/projected/f98a0de8-b0a6-4c33-83b9-831c88485e50-kube-api-access-jczn6\") pod \"redhat-operators-swvjn\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.512277 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tbfmx"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.521422 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.522437 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.551912 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tbfmx"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.566467 4858 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.566567 4858 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.566904 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.625917 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-catalog-content\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.626356 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54w7b\" (UniqueName: \"kubernetes.io/projected/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-kube-api-access-54w7b\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.626438 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-utilities\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.631169 4858 ???:1] "http: TLS handshake error from 192.168.126.11:51118: no serving certificate available for the kubelet" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.710225 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8crqq\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.722473 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.727720 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54w7b\" (UniqueName: \"kubernetes.io/projected/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-kube-api-access-54w7b\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.727787 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-catalog-content\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.727863 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-utilities\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.728722 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-utilities\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.728895 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-catalog-content\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.769044 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54w7b\" (UniqueName: \"kubernetes.io/projected/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-kube-api-access-54w7b\") pod \"redhat-operators-tbfmx\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: W0320 09:00:40.774601 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad905103_b4e5_408e_bee8_15591d245f7b.slice/crio-5f1632a838ff350bea372e92407799af515ccba055a6ddc51a539d8ce414cb27 WatchSource:0}: Error finding container 5f1632a838ff350bea372e92407799af515ccba055a6ddc51a539d8ce414cb27: Status 404 returned error can't find the container with id 5f1632a838ff350bea372e92407799af515ccba055a6ddc51a539d8ce414cb27 Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.788070 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mw84k"] Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.861647 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.877393 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.904025 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 20 09:00:40 crc kubenswrapper[4858]: W0320 09:00:40.969066 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod27e4af03_f882_4a5d_b5c6_0b8a151bb29e.slice/crio-97d1399f73359ad136e9fe02ff406689d24504ae7f08686fdab6f45fc5c0406d WatchSource:0}: Error finding container 97d1399f73359ad136e9fe02ff406689d24504ae7f08686fdab6f45fc5c0406d: Status 404 returned error can't find the container with id 97d1399f73359ad136e9fe02ff406689d24504ae7f08686fdab6f45fc5c0406d Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.996115 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" event={"ID":"ad905103-b4e5-408e-bee8-15591d245f7b","Type":"ContainerStarted","Data":"5f1632a838ff350bea372e92407799af515ccba055a6ddc51a539d8ce414cb27"} Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.997785 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mw84k" event={"ID":"57d2d9c4-3ee7-41f7-af06-18c775cb10c4","Type":"ContainerStarted","Data":"867facedf8678579b70edeb18020c2d610d9275e9618ff1ff9e6d6ea47fa0786"} Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.999738 4858 generic.go:334] "Generic (PLEG): container finished" podID="d8f37a14-0144-4154-a087-126fde1633eb" containerID="fe924a2f8c500c5a40f91694fab8c4b75b9ee5cad87fd09b14576ce594d745c0" exitCode=0 Mar 20 09:00:40 crc kubenswrapper[4858]: I0320 09:00:40.999818 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" event={"ID":"d8f37a14-0144-4154-a087-126fde1633eb","Type":"ContainerDied","Data":"fe924a2f8c500c5a40f91694fab8c4b75b9ee5cad87fd09b14576ce594d745c0"} Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.018272 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"27e4af03-f882-4a5d-b5c6-0b8a151bb29e","Type":"ContainerStarted","Data":"97d1399f73359ad136e9fe02ff406689d24504ae7f08686fdab6f45fc5c0406d"} Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.033971 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" event={"ID":"fa280af1-a47c-4359-897c-08cc65196a53","Type":"ContainerStarted","Data":"aed3956991f74001101677709e93a703a6172f718280d673b4ddc3fceecb48f4"} Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.036620 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-swvjn"] Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.036680 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.051723 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.062458 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" event={"ID":"978fa44d-6fb3-4775-ad8f-d81a582b521e","Type":"ContainerStarted","Data":"5043be7fda48d97f773179d8d583abf3abe66a313ac7abd8d4c8b5c41acac4b2"} Mar 20 09:00:41 crc kubenswrapper[4858]: W0320 09:00:41.065791 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf98a0de8_b0a6_4c33_83b9_831c88485e50.slice/crio-685c6bdfa2e9045b01c50b1b66cbeb437b492e6e213b3510b2a6570519c75dfb WatchSource:0}: Error finding container 685c6bdfa2e9045b01c50b1b66cbeb437b492e6e213b3510b2a6570519c75dfb: Status 404 returned error can't find the container with id 685c6bdfa2e9045b01c50b1b66cbeb437b492e6e213b3510b2a6570519c75dfb Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.066562 4858 generic.go:334] "Generic (PLEG): container finished" podID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerID="9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9" exitCode=0 Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.066893 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbnzw" event={"ID":"508d2d5b-0a75-4130-a396-9253b685e2cd","Type":"ContainerDied","Data":"9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9"} Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.066934 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbnzw" event={"ID":"508d2d5b-0a75-4130-a396-9253b685e2cd","Type":"ContainerStarted","Data":"2bfcb2274db4b700b51d133311f6673612f36e1ed3fdf10ace4e3fa5320ab875"} Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.084822 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" podStartSLOduration=4.084781729 podStartE2EDuration="4.084781729s" podCreationTimestamp="2026-03-20 09:00:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:41.060226389 +0000 UTC m=+222.380644576" watchObservedRunningTime="2026-03-20 09:00:41.084781729 +0000 UTC m=+222.405199926" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.154612 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-tbsv5" podStartSLOduration=15.15459135 podStartE2EDuration="15.15459135s" podCreationTimestamp="2026-03-20 09:00:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:41.14711834 +0000 UTC m=+222.467536537" watchObservedRunningTime="2026-03-20 09:00:41.15459135 +0000 UTC m=+222.475009547" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.201704 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.202822 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.206434 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.209489 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.217166 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.345907 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:41 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:41 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:41 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.346496 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.347732 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.347799 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.449803 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.449965 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.450099 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.476127 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tbfmx"] Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.506120 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: W0320 09:00:41.524455 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e6ac9fc_835d_4763_a6f5_e6923a5ee981.slice/crio-407758f9bb8bd82a3c7a4cc23b362dc44715169b1fd989abfdaabee06b7346bf WatchSource:0}: Error finding container 407758f9bb8bd82a3c7a4cc23b362dc44715169b1fd989abfdaabee06b7346bf: Status 404 returned error can't find the container with id 407758f9bb8bd82a3c7a4cc23b362dc44715169b1fd989abfdaabee06b7346bf Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.537732 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8crqq"] Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.547391 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:00:41 crc kubenswrapper[4858]: W0320 09:00:41.553817 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc2e4d497_a390_4102_961e_8334641b8867.slice/crio-262a64e95dff06f5ab981c81882b0c82fd7e84f143bc1093af68dffdde952bf9 WatchSource:0}: Error finding container 262a64e95dff06f5ab981c81882b0c82fd7e84f143bc1093af68dffdde952bf9: Status 404 returned error can't find the container with id 262a64e95dff06f5ab981c81882b0c82fd7e84f143bc1093af68dffdde952bf9 Mar 20 09:00:41 crc kubenswrapper[4858]: I0320 09:00:41.913341 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.098129 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.098751 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="effb9468-b572-4eb1-84df-15e7b0201dbf" path="/var/lib/kubelet/pods/effb9468-b572-4eb1-84df-15e7b0201dbf/volumes" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.142331 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbfmx" event={"ID":"9e6ac9fc-835d-4763-a6f5-e6923a5ee981","Type":"ContainerStarted","Data":"407758f9bb8bd82a3c7a4cc23b362dc44715169b1fd989abfdaabee06b7346bf"} Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.147912 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swvjn" event={"ID":"f98a0de8-b0a6-4c33-83b9-831c88485e50","Type":"ContainerStarted","Data":"685c6bdfa2e9045b01c50b1b66cbeb437b492e6e213b3510b2a6570519c75dfb"} Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.150673 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" event={"ID":"c2e4d497-a390-4102-961e-8334641b8867","Type":"ContainerStarted","Data":"262a64e95dff06f5ab981c81882b0c82fd7e84f143bc1093af68dffdde952bf9"} Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.155682 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" event={"ID":"ad905103-b4e5-408e-bee8-15591d245f7b","Type":"ContainerStarted","Data":"645df193bcd37a0447082bf81fe775f76a08a9d7ba3e5f6c0993f3956c0cc1da"} Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.157065 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.162559 4858 generic.go:334] "Generic (PLEG): container finished" podID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerID="5777b847f61be65de3e998433933800950cc4d41455e48965afca5581c8c4075" exitCode=0 Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.162639 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mw84k" event={"ID":"57d2d9c4-3ee7-41f7-af06-18c775cb10c4","Type":"ContainerDied","Data":"5777b847f61be65de3e998433933800950cc4d41455e48965afca5581c8c4075"} Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.162917 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.180556 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ddec1cd5-2b0d-417b-93cf-c6d76f02041c","Type":"ContainerStarted","Data":"de1ff78e4335dfdfd843b1fd05e3c414aabfeddce793972af29b4d3bac1481b5"} Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.225023 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" podStartSLOduration=5.225003254 podStartE2EDuration="5.225003254s" podCreationTimestamp="2026-03-20 09:00:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:42.200482695 +0000 UTC m=+223.520900892" watchObservedRunningTime="2026-03-20 09:00:42.225003254 +0000 UTC m=+223.545421451" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.347720 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:42 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:42 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:42 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.347786 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.670555 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.769122 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxg4l\" (UniqueName: \"kubernetes.io/projected/d8f37a14-0144-4154-a087-126fde1633eb-kube-api-access-pxg4l\") pod \"d8f37a14-0144-4154-a087-126fde1633eb\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.769535 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8f37a14-0144-4154-a087-126fde1633eb-secret-volume\") pod \"d8f37a14-0144-4154-a087-126fde1633eb\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.769571 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8f37a14-0144-4154-a087-126fde1633eb-config-volume\") pod \"d8f37a14-0144-4154-a087-126fde1633eb\" (UID: \"d8f37a14-0144-4154-a087-126fde1633eb\") " Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.770618 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8f37a14-0144-4154-a087-126fde1633eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "d8f37a14-0144-4154-a087-126fde1633eb" (UID: "d8f37a14-0144-4154-a087-126fde1633eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.783240 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8f37a14-0144-4154-a087-126fde1633eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d8f37a14-0144-4154-a087-126fde1633eb" (UID: "d8f37a14-0144-4154-a087-126fde1633eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.804301 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8f37a14-0144-4154-a087-126fde1633eb-kube-api-access-pxg4l" (OuterVolumeSpecName: "kube-api-access-pxg4l") pod "d8f37a14-0144-4154-a087-126fde1633eb" (UID: "d8f37a14-0144-4154-a087-126fde1633eb"). InnerVolumeSpecName "kube-api-access-pxg4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.871767 4858 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d8f37a14-0144-4154-a087-126fde1633eb-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.871835 4858 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d8f37a14-0144-4154-a087-126fde1633eb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:42 crc kubenswrapper[4858]: I0320 09:00:42.871870 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxg4l\" (UniqueName: \"kubernetes.io/projected/d8f37a14-0144-4154-a087-126fde1633eb-kube-api-access-pxg4l\") on node \"crc\" DevicePath \"\"" Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.203419 4858 generic.go:334] "Generic (PLEG): container finished" podID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerID="e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102" exitCode=0 Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.203571 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swvjn" event={"ID":"f98a0de8-b0a6-4c33-83b9-831c88485e50","Type":"ContainerDied","Data":"e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102"} Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.208915 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" event={"ID":"c2e4d497-a390-4102-961e-8334641b8867","Type":"ContainerStarted","Data":"11963e8bc58b0b524149f0be133e60054814a9211fbbd86007e04d09ebdc89ca"} Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.209155 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.213530 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" event={"ID":"d8f37a14-0144-4154-a087-126fde1633eb","Type":"ContainerDied","Data":"1f6ea64cacb834da961c70eae2de0ad2e965343c0d0bdb690ae7ded7e3fc0927"} Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.213562 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd" Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.213576 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f6ea64cacb834da961c70eae2de0ad2e965343c0d0bdb690ae7ded7e3fc0927" Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.217039 4858 generic.go:334] "Generic (PLEG): container finished" podID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerID="d038200a11492b4bc54f1c69bef528994645f18ce99effb5afceec12cbde6d3a" exitCode=0 Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.217138 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbfmx" event={"ID":"9e6ac9fc-835d-4763-a6f5-e6923a5ee981","Type":"ContainerDied","Data":"d038200a11492b4bc54f1c69bef528994645f18ce99effb5afceec12cbde6d3a"} Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.231149 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"27e4af03-f882-4a5d-b5c6-0b8a151bb29e","Type":"ContainerStarted","Data":"30445cbf0e2841d1663002a1b38ae0e8330a04d748e3851b079c0572980ce15f"} Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.239704 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" podStartSLOduration=168.239674696 podStartE2EDuration="2m48.239674696s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:43.234107299 +0000 UTC m=+224.554525506" watchObservedRunningTime="2026-03-20 09:00:43.239674696 +0000 UTC m=+224.560092893" Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.281527 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.28149831 podStartE2EDuration="4.28149831s" podCreationTimestamp="2026-03-20 09:00:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:00:43.277696213 +0000 UTC m=+224.598114430" watchObservedRunningTime="2026-03-20 09:00:43.28149831 +0000 UTC m=+224.601916507" Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.341667 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:43 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:43 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:43 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:43 crc kubenswrapper[4858]: I0320 09:00:43.341731 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:44 crc kubenswrapper[4858]: I0320 09:00:44.062196 4858 ???:1] "http: TLS handshake error from 192.168.126.11:37198: no serving certificate available for the kubelet" Mar 20 09:00:44 crc kubenswrapper[4858]: I0320 09:00:44.243620 4858 generic.go:334] "Generic (PLEG): container finished" podID="ddec1cd5-2b0d-417b-93cf-c6d76f02041c" containerID="d53af6191f72200166cf3ef7df4ee9f1d7a73fdd992675e40d574df001a19d6b" exitCode=0 Mar 20 09:00:44 crc kubenswrapper[4858]: I0320 09:00:44.243714 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ddec1cd5-2b0d-417b-93cf-c6d76f02041c","Type":"ContainerDied","Data":"d53af6191f72200166cf3ef7df4ee9f1d7a73fdd992675e40d574df001a19d6b"} Mar 20 09:00:44 crc kubenswrapper[4858]: I0320 09:00:44.249826 4858 generic.go:334] "Generic (PLEG): container finished" podID="27e4af03-f882-4a5d-b5c6-0b8a151bb29e" containerID="30445cbf0e2841d1663002a1b38ae0e8330a04d748e3851b079c0572980ce15f" exitCode=0 Mar 20 09:00:44 crc kubenswrapper[4858]: I0320 09:00:44.250161 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"27e4af03-f882-4a5d-b5c6-0b8a151bb29e","Type":"ContainerDied","Data":"30445cbf0e2841d1663002a1b38ae0e8330a04d748e3851b079c0572980ce15f"} Mar 20 09:00:44 crc kubenswrapper[4858]: I0320 09:00:44.335716 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:44 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:44 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:44 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:44 crc kubenswrapper[4858]: I0320 09:00:44.335792 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:45 crc kubenswrapper[4858]: I0320 09:00:45.327938 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-c47vz" Mar 20 09:00:45 crc kubenswrapper[4858]: I0320 09:00:45.335621 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:45 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:45 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:45 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:45 crc kubenswrapper[4858]: I0320 09:00:45.335728 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:46 crc kubenswrapper[4858]: I0320 09:00:46.336093 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:46 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:46 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:46 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:46 crc kubenswrapper[4858]: I0320 09:00:46.336642 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:47 crc kubenswrapper[4858]: I0320 09:00:47.336358 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:47 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:47 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:47 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:47 crc kubenswrapper[4858]: I0320 09:00:47.336463 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:48 crc kubenswrapper[4858]: I0320 09:00:48.335941 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:48 crc kubenswrapper[4858]: [-]has-synced failed: reason withheld Mar 20 09:00:48 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:48 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:48 crc kubenswrapper[4858]: I0320 09:00:48.336698 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.143510 4858 patch_prober.go:28] interesting pod/downloads-7954f5f757-j6mmm container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.143539 4858 patch_prober.go:28] interesting pod/downloads-7954f5f757-j6mmm container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.143579 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-j6mmm" podUID="e07edf68-41a8-4175-adc0-163e46620ab4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.143640 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-j6mmm" podUID="e07edf68-41a8-4175-adc0-163e46620ab4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.335114 4858 patch_prober.go:28] interesting pod/router-default-5444994796-vtwn4 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 20 09:00:49 crc kubenswrapper[4858]: [+]has-synced ok Mar 20 09:00:49 crc kubenswrapper[4858]: [+]process-running ok Mar 20 09:00:49 crc kubenswrapper[4858]: healthz check failed Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.335186 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-vtwn4" podUID="830fcf94-999e-4859-a62e-f317fc53eaf6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.454840 4858 patch_prober.go:28] interesting pod/console-f9d7485db-wr84h container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 20 09:00:49 crc kubenswrapper[4858]: I0320 09:00:49.454935 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-wr84h" podUID="91be84d3-8196-44bb-8a88-e9e6548377a1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 20 09:00:50 crc kubenswrapper[4858]: I0320 09:00:50.339123 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:50 crc kubenswrapper[4858]: I0320 09:00:50.351665 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-vtwn4" Mar 20 09:00:56 crc kubenswrapper[4858]: I0320 09:00:56.389796 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658bbb8c56-mkjhs"] Mar 20 09:00:56 crc kubenswrapper[4858]: I0320 09:00:56.400888 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" podUID="fa280af1-a47c-4359-897c-08cc65196a53" containerName="controller-manager" containerID="cri-o://aed3956991f74001101677709e93a703a6172f718280d673b4ddc3fceecb48f4" gracePeriod=30 Mar 20 09:00:56 crc kubenswrapper[4858]: I0320 09:00:56.428156 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g"] Mar 20 09:00:56 crc kubenswrapper[4858]: I0320 09:00:56.428645 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" podUID="ad905103-b4e5-408e-bee8-15591d245f7b" containerName="route-controller-manager" containerID="cri-o://645df193bcd37a0447082bf81fe775f76a08a9d7ba3e5f6c0993f3956c0cc1da" gracePeriod=30 Mar 20 09:00:58 crc kubenswrapper[4858]: I0320 09:00:58.398184 4858 generic.go:334] "Generic (PLEG): container finished" podID="fa280af1-a47c-4359-897c-08cc65196a53" containerID="aed3956991f74001101677709e93a703a6172f718280d673b4ddc3fceecb48f4" exitCode=0 Mar 20 09:00:58 crc kubenswrapper[4858]: I0320 09:00:58.398349 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" event={"ID":"fa280af1-a47c-4359-897c-08cc65196a53","Type":"ContainerDied","Data":"aed3956991f74001101677709e93a703a6172f718280d673b4ddc3fceecb48f4"} Mar 20 09:00:58 crc kubenswrapper[4858]: I0320 09:00:58.401198 4858 generic.go:334] "Generic (PLEG): container finished" podID="ad905103-b4e5-408e-bee8-15591d245f7b" containerID="645df193bcd37a0447082bf81fe775f76a08a9d7ba3e5f6c0993f3956c0cc1da" exitCode=0 Mar 20 09:00:58 crc kubenswrapper[4858]: I0320 09:00:58.401281 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" event={"ID":"ad905103-b4e5-408e-bee8-15591d245f7b","Type":"ContainerDied","Data":"645df193bcd37a0447082bf81fe775f76a08a9d7ba3e5f6c0993f3956c0cc1da"} Mar 20 09:00:58 crc kubenswrapper[4858]: I0320 09:00:58.500104 4858 patch_prober.go:28] interesting pod/controller-manager-658bbb8c56-mkjhs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" start-of-body= Mar 20 09:00:58 crc kubenswrapper[4858]: I0320 09:00:58.500223 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" podUID="fa280af1-a47c-4359-897c-08cc65196a53" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" Mar 20 09:00:59 crc kubenswrapper[4858]: I0320 09:00:59.150061 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-j6mmm" Mar 20 09:00:59 crc kubenswrapper[4858]: I0320 09:00:59.460365 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:59 crc kubenswrapper[4858]: I0320 09:00:59.465385 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:00:59 crc kubenswrapper[4858]: I0320 09:00:59.816628 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:59 crc kubenswrapper[4858]: I0320 09:00:59.837932 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/eb1ef726-a1a8-4efe-bdcc-33fba0e077ea-metrics-certs\") pod \"network-metrics-daemon-kvlch\" (UID: \"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea\") " pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:00:59 crc kubenswrapper[4858]: I0320 09:00:59.994147 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 09:01:00 crc kubenswrapper[4858]: I0320 09:01:00.001569 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-kvlch" Mar 20 09:01:00 crc kubenswrapper[4858]: I0320 09:01:00.884149 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.268944 4858 patch_prober.go:28] interesting pod/route-controller-manager-78cbcb7cb6-88g8g container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.53:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.269047 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" podUID="ad905103-b4e5-408e-bee8-15591d245f7b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.53:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.499751 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.505666 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.659513 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kube-api-access\") pod \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.659607 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kubelet-dir\") pod \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\" (UID: \"27e4af03-f882-4a5d-b5c6-0b8a151bb29e\") " Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.659659 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kube-api-access\") pod \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.659740 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "27e4af03-f882-4a5d-b5c6-0b8a151bb29e" (UID: "27e4af03-f882-4a5d-b5c6-0b8a151bb29e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.659815 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kubelet-dir\") pod \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\" (UID: \"ddec1cd5-2b0d-417b-93cf-c6d76f02041c\") " Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.659954 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ddec1cd5-2b0d-417b-93cf-c6d76f02041c" (UID: "ddec1cd5-2b0d-417b-93cf-c6d76f02041c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.660160 4858 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.660188 4858 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.667704 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "27e4af03-f882-4a5d-b5c6-0b8a151bb29e" (UID: "27e4af03-f882-4a5d-b5c6-0b8a151bb29e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.667846 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ddec1cd5-2b0d-417b-93cf-c6d76f02041c" (UID: "ddec1cd5-2b0d-417b-93cf-c6d76f02041c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.761842 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/27e4af03-f882-4a5d-b5c6-0b8a151bb29e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:01 crc kubenswrapper[4858]: I0320 09:01:01.761918 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddec1cd5-2b0d-417b-93cf-c6d76f02041c-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:02 crc kubenswrapper[4858]: I0320 09:01:02.429833 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"ddec1cd5-2b0d-417b-93cf-c6d76f02041c","Type":"ContainerDied","Data":"de1ff78e4335dfdfd843b1fd05e3c414aabfeddce793972af29b4d3bac1481b5"} Mar 20 09:01:02 crc kubenswrapper[4858]: I0320 09:01:02.434522 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1ff78e4335dfdfd843b1fd05e3c414aabfeddce793972af29b4d3bac1481b5" Mar 20 09:01:02 crc kubenswrapper[4858]: I0320 09:01:02.434575 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"27e4af03-f882-4a5d-b5c6-0b8a151bb29e","Type":"ContainerDied","Data":"97d1399f73359ad136e9fe02ff406689d24504ae7f08686fdab6f45fc5c0406d"} Mar 20 09:01:02 crc kubenswrapper[4858]: I0320 09:01:02.434607 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97d1399f73359ad136e9fe02ff406689d24504ae7f08686fdab6f45fc5c0406d" Mar 20 09:01:02 crc kubenswrapper[4858]: I0320 09:01:02.432064 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 20 09:01:02 crc kubenswrapper[4858]: I0320 09:01:02.429879 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 20 09:01:04 crc kubenswrapper[4858]: I0320 09:01:04.570605 4858 ???:1] "http: TLS handshake error from 192.168.126.11:44262: no serving certificate available for the kubelet" Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.810629 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.929717 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hhmb\" (UniqueName: \"kubernetes.io/projected/ad905103-b4e5-408e-bee8-15591d245f7b-kube-api-access-5hhmb\") pod \"ad905103-b4e5-408e-bee8-15591d245f7b\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.930174 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-config\") pod \"ad905103-b4e5-408e-bee8-15591d245f7b\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.930203 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-client-ca\") pod \"ad905103-b4e5-408e-bee8-15591d245f7b\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.930225 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad905103-b4e5-408e-bee8-15591d245f7b-serving-cert\") pod \"ad905103-b4e5-408e-bee8-15591d245f7b\" (UID: \"ad905103-b4e5-408e-bee8-15591d245f7b\") " Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.932084 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-client-ca" (OuterVolumeSpecName: "client-ca") pod "ad905103-b4e5-408e-bee8-15591d245f7b" (UID: "ad905103-b4e5-408e-bee8-15591d245f7b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.932292 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-config" (OuterVolumeSpecName: "config") pod "ad905103-b4e5-408e-bee8-15591d245f7b" (UID: "ad905103-b4e5-408e-bee8-15591d245f7b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.951044 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad905103-b4e5-408e-bee8-15591d245f7b-kube-api-access-5hhmb" (OuterVolumeSpecName: "kube-api-access-5hhmb") pod "ad905103-b4e5-408e-bee8-15591d245f7b" (UID: "ad905103-b4e5-408e-bee8-15591d245f7b"). InnerVolumeSpecName "kube-api-access-5hhmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:05 crc kubenswrapper[4858]: I0320 09:01:05.951901 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad905103-b4e5-408e-bee8-15591d245f7b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ad905103-b4e5-408e-bee8-15591d245f7b" (UID: "ad905103-b4e5-408e-bee8-15591d245f7b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.031794 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.031849 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad905103-b4e5-408e-bee8-15591d245f7b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.031872 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad905103-b4e5-408e-bee8-15591d245f7b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.031892 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hhmb\" (UniqueName: \"kubernetes.io/projected/ad905103-b4e5-408e-bee8-15591d245f7b-kube-api-access-5hhmb\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.476434 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" event={"ID":"ad905103-b4e5-408e-bee8-15591d245f7b","Type":"ContainerDied","Data":"5f1632a838ff350bea372e92407799af515ccba055a6ddc51a539d8ce414cb27"} Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.476521 4858 scope.go:117] "RemoveContainer" containerID="645df193bcd37a0447082bf81fe775f76a08a9d7ba3e5f6c0993f3956c0cc1da" Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.476663 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g" Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.496836 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g"] Mar 20 09:01:06 crc kubenswrapper[4858]: I0320 09:01:06.499463 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-78cbcb7cb6-88g8g"] Mar 20 09:01:07 crc kubenswrapper[4858]: I0320 09:01:07.898383 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:01:07 crc kubenswrapper[4858]: I0320 09:01:07.898492 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.082824 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad905103-b4e5-408e-bee8-15591d245f7b" path="/var/lib/kubelet/pods/ad905103-b4e5-408e-bee8-15591d245f7b/volumes" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092244 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4"] Mar 20 09:01:08 crc kubenswrapper[4858]: E0320 09:01:08.092647 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddec1cd5-2b0d-417b-93cf-c6d76f02041c" containerName="pruner" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092664 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddec1cd5-2b0d-417b-93cf-c6d76f02041c" containerName="pruner" Mar 20 09:01:08 crc kubenswrapper[4858]: E0320 09:01:08.092675 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8f37a14-0144-4154-a087-126fde1633eb" containerName="collect-profiles" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092681 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8f37a14-0144-4154-a087-126fde1633eb" containerName="collect-profiles" Mar 20 09:01:08 crc kubenswrapper[4858]: E0320 09:01:08.092692 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27e4af03-f882-4a5d-b5c6-0b8a151bb29e" containerName="pruner" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092702 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="27e4af03-f882-4a5d-b5c6-0b8a151bb29e" containerName="pruner" Mar 20 09:01:08 crc kubenswrapper[4858]: E0320 09:01:08.092715 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad905103-b4e5-408e-bee8-15591d245f7b" containerName="route-controller-manager" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092722 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad905103-b4e5-408e-bee8-15591d245f7b" containerName="route-controller-manager" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092816 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8f37a14-0144-4154-a087-126fde1633eb" containerName="collect-profiles" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092830 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad905103-b4e5-408e-bee8-15591d245f7b" containerName="route-controller-manager" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092840 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="27e4af03-f882-4a5d-b5c6-0b8a151bb29e" containerName="pruner" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.092847 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddec1cd5-2b0d-417b-93cf-c6d76f02041c" containerName="pruner" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.093306 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.095406 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.096232 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.096605 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.096677 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.096765 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.096900 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.108054 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4"] Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.167425 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-config\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.167484 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9625c03f-59f8-4dac-abcd-72c3bb990359-serving-cert\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.167518 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-client-ca\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.167537 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4twrw\" (UniqueName: \"kubernetes.io/projected/9625c03f-59f8-4dac-abcd-72c3bb990359-kube-api-access-4twrw\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.269004 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-client-ca\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.269060 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4twrw\" (UniqueName: \"kubernetes.io/projected/9625c03f-59f8-4dac-abcd-72c3bb990359-kube-api-access-4twrw\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.269179 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-config\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.269207 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9625c03f-59f8-4dac-abcd-72c3bb990359-serving-cert\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.271232 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-client-ca\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.271784 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-config\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.276114 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9625c03f-59f8-4dac-abcd-72c3bb990359-serving-cert\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.288138 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4twrw\" (UniqueName: \"kubernetes.io/projected/9625c03f-59f8-4dac-abcd-72c3bb990359-kube-api-access-4twrw\") pod \"route-controller-manager-56d9cd56bd-84fl4\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:08 crc kubenswrapper[4858]: I0320 09:01:08.413183 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:09 crc kubenswrapper[4858]: I0320 09:01:09.499084 4858 patch_prober.go:28] interesting pod/controller-manager-658bbb8c56-mkjhs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 20 09:01:09 crc kubenswrapper[4858]: I0320 09:01:09.499172 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" podUID="fa280af1-a47c-4359-897c-08cc65196a53" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.165619 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-qgkz2" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.691946 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.814278 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa280af1-a47c-4359-897c-08cc65196a53-serving-cert\") pod \"fa280af1-a47c-4359-897c-08cc65196a53\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.814681 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-proxy-ca-bundles\") pod \"fa280af1-a47c-4359-897c-08cc65196a53\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.814726 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-config\") pod \"fa280af1-a47c-4359-897c-08cc65196a53\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.814824 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-client-ca\") pod \"fa280af1-a47c-4359-897c-08cc65196a53\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.814956 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-977pd\" (UniqueName: \"kubernetes.io/projected/fa280af1-a47c-4359-897c-08cc65196a53-kube-api-access-977pd\") pod \"fa280af1-a47c-4359-897c-08cc65196a53\" (UID: \"fa280af1-a47c-4359-897c-08cc65196a53\") " Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.815946 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-client-ca" (OuterVolumeSpecName: "client-ca") pod "fa280af1-a47c-4359-897c-08cc65196a53" (UID: "fa280af1-a47c-4359-897c-08cc65196a53"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.816102 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-config" (OuterVolumeSpecName: "config") pod "fa280af1-a47c-4359-897c-08cc65196a53" (UID: "fa280af1-a47c-4359-897c-08cc65196a53"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.816497 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.816525 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.816742 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "fa280af1-a47c-4359-897c-08cc65196a53" (UID: "fa280af1-a47c-4359-897c-08cc65196a53"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.820020 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa280af1-a47c-4359-897c-08cc65196a53-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fa280af1-a47c-4359-897c-08cc65196a53" (UID: "fa280af1-a47c-4359-897c-08cc65196a53"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.820849 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa280af1-a47c-4359-897c-08cc65196a53-kube-api-access-977pd" (OuterVolumeSpecName: "kube-api-access-977pd") pod "fa280af1-a47c-4359-897c-08cc65196a53" (UID: "fa280af1-a47c-4359-897c-08cc65196a53"). InnerVolumeSpecName "kube-api-access-977pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.917249 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-977pd\" (UniqueName: \"kubernetes.io/projected/fa280af1-a47c-4359-897c-08cc65196a53-kube-api-access-977pd\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.917284 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa280af1-a47c-4359-897c-08cc65196a53-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:10 crc kubenswrapper[4858]: I0320 09:01:10.917296 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fa280af1-a47c-4359-897c-08cc65196a53-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:11 crc kubenswrapper[4858]: I0320 09:01:11.512251 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" event={"ID":"fa280af1-a47c-4359-897c-08cc65196a53","Type":"ContainerDied","Data":"8629c4be903257eea5eda9353696b1fb2c21fba156a98bc6cadb1fd3086c6dab"} Mar 20 09:01:11 crc kubenswrapper[4858]: I0320 09:01:11.512304 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658bbb8c56-mkjhs" Mar 20 09:01:11 crc kubenswrapper[4858]: I0320 09:01:11.544066 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658bbb8c56-mkjhs"] Mar 20 09:01:11 crc kubenswrapper[4858]: I0320 09:01:11.547496 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-658bbb8c56-mkjhs"] Mar 20 09:01:11 crc kubenswrapper[4858]: I0320 09:01:11.578227 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.080002 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa280af1-a47c-4359-897c-08cc65196a53" path="/var/lib/kubelet/pods/fa280af1-a47c-4359-897c-08cc65196a53/volumes" Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.985217 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 09:01:12 crc kubenswrapper[4858]: E0320 09:01:12.985599 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa280af1-a47c-4359-897c-08cc65196a53" containerName="controller-manager" Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.985629 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa280af1-a47c-4359-897c-08cc65196a53" containerName="controller-manager" Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.985794 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa280af1-a47c-4359-897c-08cc65196a53" containerName="controller-manager" Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.986504 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.992753 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.993338 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 20 09:01:12 crc kubenswrapper[4858]: I0320 09:01:12.993757 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 20 09:01:13 crc kubenswrapper[4858]: I0320 09:01:13.152090 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61c99e8c-3178-4986-af83-38beb9284875-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:13 crc kubenswrapper[4858]: I0320 09:01:13.152202 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c99e8c-3178-4986-af83-38beb9284875-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:13 crc kubenswrapper[4858]: I0320 09:01:13.254268 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61c99e8c-3178-4986-af83-38beb9284875-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:13 crc kubenswrapper[4858]: I0320 09:01:13.254402 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61c99e8c-3178-4986-af83-38beb9284875-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:13 crc kubenswrapper[4858]: I0320 09:01:13.254493 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c99e8c-3178-4986-af83-38beb9284875-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:13 crc kubenswrapper[4858]: I0320 09:01:13.285999 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c99e8c-3178-4986-af83-38beb9284875-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:13 crc kubenswrapper[4858]: I0320 09:01:13.320947 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:15 crc kubenswrapper[4858]: E0320 09:01:15.543635 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 20 09:01:15 crc kubenswrapper[4858]: E0320 09:01:15.544025 4858 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 20 09:01:15 crc kubenswrapper[4858]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 20 09:01:15 crc kubenswrapper[4858]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v6scs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29566620-gz4tx_openshift-infra(74fe10ec-a162-4c93-b2d3-1a80745e7fcc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 20 09:01:15 crc kubenswrapper[4858]: > logger="UnhandledError" Mar 20 09:01:15 crc kubenswrapper[4858]: E0320 09:01:15.545218 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" podUID="74fe10ec-a162-4c93-b2d3-1a80745e7fcc" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.105703 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-585774d4-cz2nv"] Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.107173 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.110835 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.110954 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.111217 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.111427 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.111615 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.111860 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-585774d4-cz2nv"] Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.111965 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.120970 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.204285 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sblx\" (UniqueName: \"kubernetes.io/projected/d7df7e6e-405b-451d-8955-ff30d26c368a-kube-api-access-6sblx\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.204609 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-config\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.204644 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7df7e6e-405b-451d-8955-ff30d26c368a-serving-cert\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.204699 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-proxy-ca-bundles\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.204809 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-client-ca\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.306375 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-client-ca\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.307974 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-client-ca\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.307007 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sblx\" (UniqueName: \"kubernetes.io/projected/d7df7e6e-405b-451d-8955-ff30d26c368a-kube-api-access-6sblx\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.309428 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-config\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.309463 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7df7e6e-405b-451d-8955-ff30d26c368a-serving-cert\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.309530 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-proxy-ca-bundles\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.311199 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-proxy-ca-bundles\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.311596 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-config\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.318336 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7df7e6e-405b-451d-8955-ff30d26c368a-serving-cert\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.328653 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sblx\" (UniqueName: \"kubernetes.io/projected/d7df7e6e-405b-451d-8955-ff30d26c368a-kube-api-access-6sblx\") pod \"controller-manager-585774d4-cz2nv\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.358393 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-585774d4-cz2nv"] Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.358948 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:16 crc kubenswrapper[4858]: I0320 09:01:16.462175 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4"] Mar 20 09:01:16 crc kubenswrapper[4858]: E0320 09:01:16.544820 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" podUID="74fe10ec-a162-4c93-b2d3-1a80745e7fcc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.380495 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.381643 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.385910 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.531026 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-var-lock\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.531190 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kube-api-access\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.531253 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.632613 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kube-api-access\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.632710 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.632750 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-var-lock\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.632894 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-var-lock\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.632891 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.653425 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kube-api-access\") pod \"installer-9-crc\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:17 crc kubenswrapper[4858]: I0320 09:01:17.713361 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:01:19 crc kubenswrapper[4858]: E0320 09:01:19.760507 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 20 09:01:19 crc kubenswrapper[4858]: E0320 09:01:19.761343 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-54w7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-tbfmx_openshift-marketplace(9e6ac9fc-835d-4763-a6f5-e6923a5ee981): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:19 crc kubenswrapper[4858]: E0320 09:01:19.762717 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-tbfmx" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" Mar 20 09:01:21 crc kubenswrapper[4858]: E0320 09:01:21.633168 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-tbfmx" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" Mar 20 09:01:21 crc kubenswrapper[4858]: E0320 09:01:21.706175 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 09:01:21 crc kubenswrapper[4858]: E0320 09:01:21.706386 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n9gqc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-crt7w_openshift-marketplace(c8456e28-cc53-4820-8bbf-44e27de1dc9b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:21 crc kubenswrapper[4858]: E0320 09:01:21.708129 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-crt7w" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.125739 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-crt7w" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.200821 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.201021 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6pggl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-pbnzw_openshift-marketplace(508d2d5b-0a75-4130-a396-9253b685e2cd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.202229 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-pbnzw" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.223589 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.223813 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jczn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-swvjn_openshift-marketplace(f98a0de8-b0a6-4c33-83b9-831c88485e50): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.225103 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-swvjn" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.238735 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.238978 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jqfcc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2dv2r_openshift-marketplace(e78c3dad-ee9d-4901-8c08-2db4bd2070cd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:23 crc kubenswrapper[4858]: E0320 09:01:23.240267 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2dv2r" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.748549 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-pbnzw" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.748657 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2dv2r" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" Mar 20 09:01:24 crc kubenswrapper[4858]: I0320 09:01:24.764157 4858 scope.go:117] "RemoveContainer" containerID="aed3956991f74001101677709e93a703a6172f718280d673b4ddc3fceecb48f4" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.838598 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.839391 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fll4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6kdlf_openshift-marketplace(9daba85d-2681-4f74-8094-9db79d723cee): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.841064 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6kdlf" podUID="9daba85d-2681-4f74-8094-9db79d723cee" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.847204 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.847461 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g575q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4dwft_openshift-marketplace(1b56366e-866a-4139-9b65-3228c5f92d4a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.849126 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4dwft" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.856400 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.856573 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p4jf2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mw84k_openshift-marketplace(57d2d9c4-3ee7-41f7-af06-18c775cb10c4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 20 09:01:24 crc kubenswrapper[4858]: E0320 09:01:24.858088 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mw84k" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.226186 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-585774d4-cz2nv"] Mar 20 09:01:25 crc kubenswrapper[4858]: W0320 09:01:25.233623 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7df7e6e_405b_451d_8955_ff30d26c368a.slice/crio-c0e5654ee527f78da91f19fcf46e90db2a5cdb7959b73daecf61c9382084b8ec WatchSource:0}: Error finding container c0e5654ee527f78da91f19fcf46e90db2a5cdb7959b73daecf61c9382084b8ec: Status 404 returned error can't find the container with id c0e5654ee527f78da91f19fcf46e90db2a5cdb7959b73daecf61c9382084b8ec Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.286576 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-kvlch"] Mar 20 09:01:25 crc kubenswrapper[4858]: W0320 09:01:25.288074 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb1ef726_a1a8_4efe_bdcc_33fba0e077ea.slice/crio-069490d1a7043a59775f10426e8560cd95a14d966c822179d3536b7ef78f05d5 WatchSource:0}: Error finding container 069490d1a7043a59775f10426e8560cd95a14d966c822179d3536b7ef78f05d5: Status 404 returned error can't find the container with id 069490d1a7043a59775f10426e8560cd95a14d966c822179d3536b7ef78f05d5 Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.289448 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4"] Mar 20 09:01:25 crc kubenswrapper[4858]: W0320 09:01:25.290362 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9625c03f_59f8_4dac_abcd_72c3bb990359.slice/crio-570dd5cf3bafaa8fc714dedd21fa33cf9a7ea44c32094cae881b2f5f97552ffa WatchSource:0}: Error finding container 570dd5cf3bafaa8fc714dedd21fa33cf9a7ea44c32094cae881b2f5f97552ffa: Status 404 returned error can't find the container with id 570dd5cf3bafaa8fc714dedd21fa33cf9a7ea44c32094cae881b2f5f97552ffa Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.382782 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.398257 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 20 09:01:25 crc kubenswrapper[4858]: W0320 09:01:25.413558 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod61c99e8c_3178_4986_af83_38beb9284875.slice/crio-8dc6f480d9378b6003152c26bf86f7980e7f9d04878eb5204673eba7b6e68008 WatchSource:0}: Error finding container 8dc6f480d9378b6003152c26bf86f7980e7f9d04878eb5204673eba7b6e68008: Status 404 returned error can't find the container with id 8dc6f480d9378b6003152c26bf86f7980e7f9d04878eb5204673eba7b6e68008 Mar 20 09:01:25 crc kubenswrapper[4858]: W0320 09:01:25.418685 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb42b85d2_a66d_40be_af75_611eaa9d1a3a.slice/crio-48190b4afb2273f4f4f21d5680c1722b840cbc74d8ec5340da4b421203d71503 WatchSource:0}: Error finding container 48190b4afb2273f4f4f21d5680c1722b840cbc74d8ec5340da4b421203d71503: Status 404 returned error can't find the container with id 48190b4afb2273f4f4f21d5680c1722b840cbc74d8ec5340da4b421203d71503 Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.609697 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" event={"ID":"d7df7e6e-405b-451d-8955-ff30d26c368a","Type":"ContainerStarted","Data":"12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0"} Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.609778 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" event={"ID":"d7df7e6e-405b-451d-8955-ff30d26c368a","Type":"ContainerStarted","Data":"c0e5654ee527f78da91f19fcf46e90db2a5cdb7959b73daecf61c9382084b8ec"} Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.609970 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" podUID="d7df7e6e-405b-451d-8955-ff30d26c368a" containerName="controller-manager" containerID="cri-o://12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0" gracePeriod=30 Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.611529 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.618755 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"61c99e8c-3178-4986-af83-38beb9284875","Type":"ContainerStarted","Data":"8dc6f480d9378b6003152c26bf86f7980e7f9d04878eb5204673eba7b6e68008"} Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.634109 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.637769 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" podStartSLOduration=29.637711104 podStartE2EDuration="29.637711104s" podCreationTimestamp="2026-03-20 09:00:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:25.634481405 +0000 UTC m=+266.954899602" watchObservedRunningTime="2026-03-20 09:01:25.637711104 +0000 UTC m=+266.958129301" Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.644137 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" podUID="9625c03f-59f8-4dac-abcd-72c3bb990359" containerName="route-controller-manager" containerID="cri-o://f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9" gracePeriod=30 Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.644808 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" event={"ID":"9625c03f-59f8-4dac-abcd-72c3bb990359","Type":"ContainerStarted","Data":"f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9"} Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.645029 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.645129 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" event={"ID":"9625c03f-59f8-4dac-abcd-72c3bb990359","Type":"ContainerStarted","Data":"570dd5cf3bafaa8fc714dedd21fa33cf9a7ea44c32094cae881b2f5f97552ffa"} Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.655701 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b42b85d2-a66d-40be-af75-611eaa9d1a3a","Type":"ContainerStarted","Data":"48190b4afb2273f4f4f21d5680c1722b840cbc74d8ec5340da4b421203d71503"} Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.672804 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kvlch" event={"ID":"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea","Type":"ContainerStarted","Data":"1f4276f57b1efa328730ccd767d28fab7eb7d2459ce40d4beedb1480e902accc"} Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.672862 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kvlch" event={"ID":"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea","Type":"ContainerStarted","Data":"069490d1a7043a59775f10426e8560cd95a14d966c822179d3536b7ef78f05d5"} Mar 20 09:01:25 crc kubenswrapper[4858]: E0320 09:01:25.687614 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6kdlf" podUID="9daba85d-2681-4f74-8094-9db79d723cee" Mar 20 09:01:25 crc kubenswrapper[4858]: I0320 09:01:25.701723 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" podStartSLOduration=29.701696455 podStartE2EDuration="29.701696455s" podCreationTimestamp="2026-03-20 09:00:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:25.696401226 +0000 UTC m=+267.016819423" watchObservedRunningTime="2026-03-20 09:01:25.701696455 +0000 UTC m=+267.022114662" Mar 20 09:01:25 crc kubenswrapper[4858]: E0320 09:01:25.714770 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4dwft" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" Mar 20 09:01:25 crc kubenswrapper[4858]: E0320 09:01:25.715044 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mw84k" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.070799 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.080946 4858 patch_prober.go:28] interesting pod/route-controller-manager-56d9cd56bd-84fl4 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": read tcp 10.217.0.2:60262->10.217.0.57:8443: read: connection reset by peer" start-of-body= Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.081023 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" podUID="9625c03f-59f8-4dac-abcd-72c3bb990359" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": read tcp 10.217.0.2:60262->10.217.0.57:8443: read: connection reset by peer" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.103554 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr"] Mar 20 09:01:26 crc kubenswrapper[4858]: E0320 09:01:26.103848 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7df7e6e-405b-451d-8955-ff30d26c368a" containerName="controller-manager" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.103863 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7df7e6e-405b-451d-8955-ff30d26c368a" containerName="controller-manager" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.103978 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7df7e6e-405b-451d-8955-ff30d26c368a" containerName="controller-manager" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.104483 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.120024 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr"] Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.187057 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-config\") pod \"d7df7e6e-405b-451d-8955-ff30d26c368a\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.187787 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-client-ca\") pod \"d7df7e6e-405b-451d-8955-ff30d26c368a\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.187840 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-proxy-ca-bundles\") pod \"d7df7e6e-405b-451d-8955-ff30d26c368a\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188135 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sblx\" (UniqueName: \"kubernetes.io/projected/d7df7e6e-405b-451d-8955-ff30d26c368a-kube-api-access-6sblx\") pod \"d7df7e6e-405b-451d-8955-ff30d26c368a\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188246 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7df7e6e-405b-451d-8955-ff30d26c368a-serving-cert\") pod \"d7df7e6e-405b-451d-8955-ff30d26c368a\" (UID: \"d7df7e6e-405b-451d-8955-ff30d26c368a\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188496 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d7df7e6e-405b-451d-8955-ff30d26c368a" (UID: "d7df7e6e-405b-451d-8955-ff30d26c368a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188652 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-client-ca" (OuterVolumeSpecName: "client-ca") pod "d7df7e6e-405b-451d-8955-ff30d26c368a" (UID: "d7df7e6e-405b-451d-8955-ff30d26c368a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188741 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-config" (OuterVolumeSpecName: "config") pod "d7df7e6e-405b-451d-8955-ff30d26c368a" (UID: "d7df7e6e-405b-451d-8955-ff30d26c368a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188776 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-config\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188830 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-serving-cert\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188920 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-client-ca\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188962 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-proxy-ca-bundles\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.188997 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzw5m\" (UniqueName: \"kubernetes.io/projected/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-kube-api-access-nzw5m\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.189163 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.189192 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.189208 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d7df7e6e-405b-451d-8955-ff30d26c368a-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.195588 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7df7e6e-405b-451d-8955-ff30d26c368a-kube-api-access-6sblx" (OuterVolumeSpecName: "kube-api-access-6sblx") pod "d7df7e6e-405b-451d-8955-ff30d26c368a" (UID: "d7df7e6e-405b-451d-8955-ff30d26c368a"). InnerVolumeSpecName "kube-api-access-6sblx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.198517 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7df7e6e-405b-451d-8955-ff30d26c368a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d7df7e6e-405b-451d-8955-ff30d26c368a" (UID: "d7df7e6e-405b-451d-8955-ff30d26c368a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.290339 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-config\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.290386 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-serving-cert\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.290418 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-client-ca\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.290437 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-proxy-ca-bundles\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.290530 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzw5m\" (UniqueName: \"kubernetes.io/projected/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-kube-api-access-nzw5m\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.290614 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sblx\" (UniqueName: \"kubernetes.io/projected/d7df7e6e-405b-451d-8955-ff30d26c368a-kube-api-access-6sblx\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.290626 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7df7e6e-405b-451d-8955-ff30d26c368a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.291758 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-client-ca\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.292068 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-proxy-ca-bundles\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.292072 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-config\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.295821 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-serving-cert\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.307162 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzw5m\" (UniqueName: \"kubernetes.io/projected/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-kube-api-access-nzw5m\") pod \"controller-manager-7c569f8fcc-6pzmr\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.359884 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-56d9cd56bd-84fl4_9625c03f-59f8-4dac-abcd-72c3bb990359/route-controller-manager/0.log" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.359992 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.486153 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.493699 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9625c03f-59f8-4dac-abcd-72c3bb990359-serving-cert\") pod \"9625c03f-59f8-4dac-abcd-72c3bb990359\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.493864 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4twrw\" (UniqueName: \"kubernetes.io/projected/9625c03f-59f8-4dac-abcd-72c3bb990359-kube-api-access-4twrw\") pod \"9625c03f-59f8-4dac-abcd-72c3bb990359\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.494001 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-config\") pod \"9625c03f-59f8-4dac-abcd-72c3bb990359\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.494129 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-client-ca\") pod \"9625c03f-59f8-4dac-abcd-72c3bb990359\" (UID: \"9625c03f-59f8-4dac-abcd-72c3bb990359\") " Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.494771 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-client-ca" (OuterVolumeSpecName: "client-ca") pod "9625c03f-59f8-4dac-abcd-72c3bb990359" (UID: "9625c03f-59f8-4dac-abcd-72c3bb990359"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.494801 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-config" (OuterVolumeSpecName: "config") pod "9625c03f-59f8-4dac-abcd-72c3bb990359" (UID: "9625c03f-59f8-4dac-abcd-72c3bb990359"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.499566 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9625c03f-59f8-4dac-abcd-72c3bb990359-kube-api-access-4twrw" (OuterVolumeSpecName: "kube-api-access-4twrw") pod "9625c03f-59f8-4dac-abcd-72c3bb990359" (UID: "9625c03f-59f8-4dac-abcd-72c3bb990359"). InnerVolumeSpecName "kube-api-access-4twrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.499579 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625c03f-59f8-4dac-abcd-72c3bb990359-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9625c03f-59f8-4dac-abcd-72c3bb990359" (UID: "9625c03f-59f8-4dac-abcd-72c3bb990359"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.596388 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.596420 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9625c03f-59f8-4dac-abcd-72c3bb990359-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.596430 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4twrw\" (UniqueName: \"kubernetes.io/projected/9625c03f-59f8-4dac-abcd-72c3bb990359-kube-api-access-4twrw\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.596441 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9625c03f-59f8-4dac-abcd-72c3bb990359-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.693580 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-kvlch" event={"ID":"eb1ef726-a1a8-4efe-bdcc-33fba0e077ea","Type":"ContainerStarted","Data":"b2bb4011a2c09a12202b544dfc4da9435ea5c9b39d60f7391e460b3e72d60c6f"} Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.696443 4858 generic.go:334] "Generic (PLEG): container finished" podID="d7df7e6e-405b-451d-8955-ff30d26c368a" containerID="12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0" exitCode=0 Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.696521 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" event={"ID":"d7df7e6e-405b-451d-8955-ff30d26c368a","Type":"ContainerDied","Data":"12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0"} Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.696550 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.696583 4858 scope.go:117] "RemoveContainer" containerID="12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.696568 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-585774d4-cz2nv" event={"ID":"d7df7e6e-405b-451d-8955-ff30d26c368a","Type":"ContainerDied","Data":"c0e5654ee527f78da91f19fcf46e90db2a5cdb7959b73daecf61c9382084b8ec"} Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.700724 4858 generic.go:334] "Generic (PLEG): container finished" podID="61c99e8c-3178-4986-af83-38beb9284875" containerID="321f0405c95fda84ef045dacaf3431b4df5de774de93ef543662976b80168a07" exitCode=0 Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.700797 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"61c99e8c-3178-4986-af83-38beb9284875","Type":"ContainerDied","Data":"321f0405c95fda84ef045dacaf3431b4df5de774de93ef543662976b80168a07"} Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.703703 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-56d9cd56bd-84fl4_9625c03f-59f8-4dac-abcd-72c3bb990359/route-controller-manager/0.log" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.703746 4858 generic.go:334] "Generic (PLEG): container finished" podID="9625c03f-59f8-4dac-abcd-72c3bb990359" containerID="f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9" exitCode=255 Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.703795 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" event={"ID":"9625c03f-59f8-4dac-abcd-72c3bb990359","Type":"ContainerDied","Data":"f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9"} Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.703818 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" event={"ID":"9625c03f-59f8-4dac-abcd-72c3bb990359","Type":"ContainerDied","Data":"570dd5cf3bafaa8fc714dedd21fa33cf9a7ea44c32094cae881b2f5f97552ffa"} Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.703944 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.707872 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b42b85d2-a66d-40be-af75-611eaa9d1a3a","Type":"ContainerStarted","Data":"4ca8c3ddec56c38859968f4560a7a3d304b47b5d70d59813d00b4ba6a528155d"} Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.712139 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-kvlch" podStartSLOduration=211.712113954 podStartE2EDuration="3m31.712113954s" podCreationTimestamp="2026-03-20 08:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:26.711306454 +0000 UTC m=+268.031724651" watchObservedRunningTime="2026-03-20 09:01:26.712113954 +0000 UTC m=+268.032532161" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.744775 4858 scope.go:117] "RemoveContainer" containerID="12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0" Mar 20 09:01:26 crc kubenswrapper[4858]: E0320 09:01:26.745380 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0\": container with ID starting with 12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0 not found: ID does not exist" containerID="12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.745426 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0"} err="failed to get container status \"12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0\": rpc error: code = NotFound desc = could not find container \"12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0\": container with ID starting with 12e66e94e6467a989a816b76173648babcec9cfe89cca33099069e21b8a047c0 not found: ID does not exist" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.745458 4858 scope.go:117] "RemoveContainer" containerID="f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.759597 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=9.759577932 podStartE2EDuration="9.759577932s" podCreationTimestamp="2026-03-20 09:01:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:26.759521221 +0000 UTC m=+268.079939438" watchObservedRunningTime="2026-03-20 09:01:26.759577932 +0000 UTC m=+268.079996129" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.774099 4858 scope.go:117] "RemoveContainer" containerID="f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9" Mar 20 09:01:26 crc kubenswrapper[4858]: E0320 09:01:26.776639 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9\": container with ID starting with f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9 not found: ID does not exist" containerID="f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.776716 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9"} err="failed to get container status \"f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9\": rpc error: code = NotFound desc = could not find container \"f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9\": container with ID starting with f7c05279c320156c1b546d9ac307c685f1063e6aaa4a0e9d76a9f2ac22066de9 not found: ID does not exist" Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.780358 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-585774d4-cz2nv"] Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.786489 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-585774d4-cz2nv"] Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.816297 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4"] Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.821119 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56d9cd56bd-84fl4"] Mar 20 09:01:26 crc kubenswrapper[4858]: I0320 09:01:26.899847 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr"] Mar 20 09:01:27 crc kubenswrapper[4858]: I0320 09:01:27.734144 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" event={"ID":"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67","Type":"ContainerStarted","Data":"9afb21abc17b156fb1141a5311a1806842df605aca844698373f63864ed75bec"} Mar 20 09:01:27 crc kubenswrapper[4858]: I0320 09:01:27.735721 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:27 crc kubenswrapper[4858]: I0320 09:01:27.735794 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" event={"ID":"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67","Type":"ContainerStarted","Data":"20f9fe0648a37b2bdc0698da559206b4abefab1c0fd124f5089588d63d9211e1"} Mar 20 09:01:27 crc kubenswrapper[4858]: I0320 09:01:27.768232 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" podStartSLOduration=11.768198557 podStartE2EDuration="11.768198557s" podCreationTimestamp="2026-03-20 09:01:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:27.76548666 +0000 UTC m=+269.085904867" watchObservedRunningTime="2026-03-20 09:01:27.768198557 +0000 UTC m=+269.088616774" Mar 20 09:01:27 crc kubenswrapper[4858]: I0320 09:01:27.777363 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.066808 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.103248 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9625c03f-59f8-4dac-abcd-72c3bb990359" path="/var/lib/kubelet/pods/9625c03f-59f8-4dac-abcd-72c3bb990359/volumes" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.104478 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7df7e6e-405b-451d-8955-ff30d26c368a" path="/var/lib/kubelet/pods/d7df7e6e-405b-451d-8955-ff30d26c368a/volumes" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.111047 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7"] Mar 20 09:01:28 crc kubenswrapper[4858]: E0320 09:01:28.111721 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9625c03f-59f8-4dac-abcd-72c3bb990359" containerName="route-controller-manager" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.111759 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9625c03f-59f8-4dac-abcd-72c3bb990359" containerName="route-controller-manager" Mar 20 09:01:28 crc kubenswrapper[4858]: E0320 09:01:28.111804 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c99e8c-3178-4986-af83-38beb9284875" containerName="pruner" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.111814 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c99e8c-3178-4986-af83-38beb9284875" containerName="pruner" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.112031 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c99e8c-3178-4986-af83-38beb9284875" containerName="pruner" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.112065 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="9625c03f-59f8-4dac-abcd-72c3bb990359" containerName="route-controller-manager" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.112809 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.117681 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.117734 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.118016 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.118173 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.119326 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.119564 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.128432 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61c99e8c-3178-4986-af83-38beb9284875-kubelet-dir\") pod \"61c99e8c-3178-4986-af83-38beb9284875\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.128614 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c99e8c-3178-4986-af83-38beb9284875-kube-api-access\") pod \"61c99e8c-3178-4986-af83-38beb9284875\" (UID: \"61c99e8c-3178-4986-af83-38beb9284875\") " Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.129243 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61c99e8c-3178-4986-af83-38beb9284875-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "61c99e8c-3178-4986-af83-38beb9284875" (UID: "61c99e8c-3178-4986-af83-38beb9284875"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.129631 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7"] Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.138867 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c99e8c-3178-4986-af83-38beb9284875-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "61c99e8c-3178-4986-af83-38beb9284875" (UID: "61c99e8c-3178-4986-af83-38beb9284875"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.230636 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-config\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.230762 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-serving-cert\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.230795 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dzsn\" (UniqueName: \"kubernetes.io/projected/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-kube-api-access-8dzsn\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.230820 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-client-ca\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.230859 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c99e8c-3178-4986-af83-38beb9284875-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.230874 4858 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/61c99e8c-3178-4986-af83-38beb9284875-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.332818 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-serving-cert\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.333817 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dzsn\" (UniqueName: \"kubernetes.io/projected/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-kube-api-access-8dzsn\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.333951 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-client-ca\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.334074 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-config\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.335614 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-client-ca\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.335759 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-config\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.339776 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-serving-cert\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.353811 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dzsn\" (UniqueName: \"kubernetes.io/projected/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-kube-api-access-8dzsn\") pod \"route-controller-manager-77f575b587-swmq7\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.434098 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.644565 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7"] Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.744248 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" event={"ID":"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6","Type":"ContainerStarted","Data":"7c3480dd635356c36be92ece9597e21b8181a6c9a18d88da279cd35ff99876ff"} Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.746439 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.746424 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"61c99e8c-3178-4986-af83-38beb9284875","Type":"ContainerDied","Data":"8dc6f480d9378b6003152c26bf86f7980e7f9d04878eb5204673eba7b6e68008"} Mar 20 09:01:28 crc kubenswrapper[4858]: I0320 09:01:28.747018 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dc6f480d9378b6003152c26bf86f7980e7f9d04878eb5204673eba7b6e68008" Mar 20 09:01:29 crc kubenswrapper[4858]: I0320 09:01:29.755421 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" event={"ID":"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6","Type":"ContainerStarted","Data":"c3df189adb4d3b6bd5e784894ec6e2d7aef021b7943893f069781970a606f06d"} Mar 20 09:01:29 crc kubenswrapper[4858]: I0320 09:01:29.755803 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:29 crc kubenswrapper[4858]: I0320 09:01:29.763542 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:29 crc kubenswrapper[4858]: I0320 09:01:29.783370 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" podStartSLOduration=13.783339163 podStartE2EDuration="13.783339163s" podCreationTimestamp="2026-03-20 09:01:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:29.776815634 +0000 UTC m=+271.097233851" watchObservedRunningTime="2026-03-20 09:01:29.783339163 +0000 UTC m=+271.103757380" Mar 20 09:01:33 crc kubenswrapper[4858]: I0320 09:01:33.526583 4858 csr.go:261] certificate signing request csr-x5mkj is approved, waiting to be issued Mar 20 09:01:33 crc kubenswrapper[4858]: I0320 09:01:33.534219 4858 csr.go:257] certificate signing request csr-x5mkj is issued Mar 20 09:01:33 crc kubenswrapper[4858]: I0320 09:01:33.782354 4858 generic.go:334] "Generic (PLEG): container finished" podID="74fe10ec-a162-4c93-b2d3-1a80745e7fcc" containerID="28ba9dd114777b6309024ecd2a663cdd2b72f0255cf4c838204de1a5503439e8" exitCode=0 Mar 20 09:01:33 crc kubenswrapper[4858]: I0320 09:01:33.782488 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" event={"ID":"74fe10ec-a162-4c93-b2d3-1a80745e7fcc","Type":"ContainerDied","Data":"28ba9dd114777b6309024ecd2a663cdd2b72f0255cf4c838204de1a5503439e8"} Mar 20 09:01:34 crc kubenswrapper[4858]: I0320 09:01:34.535587 4858 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-20 18:53:38.96011069 +0000 UTC Mar 20 09:01:34 crc kubenswrapper[4858]: I0320 09:01:34.535650 4858 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5889h52m4.424463131s for next certificate rotation Mar 20 09:01:35 crc kubenswrapper[4858]: I0320 09:01:35.147501 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" Mar 20 09:01:35 crc kubenswrapper[4858]: I0320 09:01:35.250584 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6scs\" (UniqueName: \"kubernetes.io/projected/74fe10ec-a162-4c93-b2d3-1a80745e7fcc-kube-api-access-v6scs\") pod \"74fe10ec-a162-4c93-b2d3-1a80745e7fcc\" (UID: \"74fe10ec-a162-4c93-b2d3-1a80745e7fcc\") " Mar 20 09:01:35 crc kubenswrapper[4858]: I0320 09:01:35.259210 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74fe10ec-a162-4c93-b2d3-1a80745e7fcc-kube-api-access-v6scs" (OuterVolumeSpecName: "kube-api-access-v6scs") pod "74fe10ec-a162-4c93-b2d3-1a80745e7fcc" (UID: "74fe10ec-a162-4c93-b2d3-1a80745e7fcc"). InnerVolumeSpecName "kube-api-access-v6scs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:35 crc kubenswrapper[4858]: I0320 09:01:35.353293 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6scs\" (UniqueName: \"kubernetes.io/projected/74fe10ec-a162-4c93-b2d3-1a80745e7fcc-kube-api-access-v6scs\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:35 crc kubenswrapper[4858]: I0320 09:01:35.797900 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" event={"ID":"74fe10ec-a162-4c93-b2d3-1a80745e7fcc","Type":"ContainerDied","Data":"5c8d754c20df28d8c41a4c9f38670efb315b937f17d8c6c667779683b96d0d23"} Mar 20 09:01:35 crc kubenswrapper[4858]: I0320 09:01:35.797956 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c8d754c20df28d8c41a4c9f38670efb315b937f17d8c6c667779683b96d0d23" Mar 20 09:01:35 crc kubenswrapper[4858]: I0320 09:01:35.797962 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566620-gz4tx" Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.347395 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr"] Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.348261 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" podUID="e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" containerName="controller-manager" containerID="cri-o://9afb21abc17b156fb1141a5311a1806842df605aca844698373f63864ed75bec" gracePeriod=30 Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.383261 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7"] Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.383688 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" podUID="6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" containerName="route-controller-manager" containerID="cri-o://c3df189adb4d3b6bd5e784894ec6e2d7aef021b7943893f069781970a606f06d" gracePeriod=30 Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.489104 4858 patch_prober.go:28] interesting pod/controller-manager-7c569f8fcc-6pzmr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" start-of-body= Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.489193 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" podUID="e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.813835 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbfmx" event={"ID":"9e6ac9fc-835d-4763-a6f5-e6923a5ee981","Type":"ContainerStarted","Data":"9dd033e13afafac720728adc96412bc8c9145a7837a885562bb5ccd265331652"} Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.823278 4858 generic.go:334] "Generic (PLEG): container finished" podID="6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" containerID="c3df189adb4d3b6bd5e784894ec6e2d7aef021b7943893f069781970a606f06d" exitCode=0 Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.823390 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" event={"ID":"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6","Type":"ContainerDied","Data":"c3df189adb4d3b6bd5e784894ec6e2d7aef021b7943893f069781970a606f06d"} Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.828483 4858 generic.go:334] "Generic (PLEG): container finished" podID="e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" containerID="9afb21abc17b156fb1141a5311a1806842df605aca844698373f63864ed75bec" exitCode=0 Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.828547 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" event={"ID":"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67","Type":"ContainerDied","Data":"9afb21abc17b156fb1141a5311a1806842df605aca844698373f63864ed75bec"} Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.949114 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:36 crc kubenswrapper[4858]: I0320 09:01:36.963937 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.077846 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-serving-cert\") pod \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.077893 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-client-ca\") pod \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.077912 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-config\") pod \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.077991 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-proxy-ca-bundles\") pod \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.078049 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dzsn\" (UniqueName: \"kubernetes.io/projected/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-kube-api-access-8dzsn\") pod \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.078077 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzw5m\" (UniqueName: \"kubernetes.io/projected/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-kube-api-access-nzw5m\") pod \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.078117 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-client-ca\") pod \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.078140 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-config\") pod \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\" (UID: \"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.078201 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-serving-cert\") pod \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\" (UID: \"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6\") " Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.080723 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-config" (OuterVolumeSpecName: "config") pod "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" (UID: "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.081161 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-client-ca" (OuterVolumeSpecName: "client-ca") pod "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" (UID: "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.081651 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" (UID: "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.081976 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-config" (OuterVolumeSpecName: "config") pod "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" (UID: "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.082093 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-client-ca" (OuterVolumeSpecName: "client-ca") pod "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" (UID: "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.086832 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-kube-api-access-8dzsn" (OuterVolumeSpecName: "kube-api-access-8dzsn") pod "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" (UID: "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6"). InnerVolumeSpecName "kube-api-access-8dzsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.086947 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" (UID: "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.087478 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" (UID: "6b144d61-bc46-4b54-9bd4-7b2127c7ffc6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.087713 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-kube-api-access-nzw5m" (OuterVolumeSpecName: "kube-api-access-nzw5m") pod "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" (UID: "e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67"). InnerVolumeSpecName "kube-api-access-nzw5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180350 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180392 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180492 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180504 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180513 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180522 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180530 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180542 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dzsn\" (UniqueName: \"kubernetes.io/projected/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6-kube-api-access-8dzsn\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.180551 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzw5m\" (UniqueName: \"kubernetes.io/projected/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67-kube-api-access-nzw5m\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.835247 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" event={"ID":"e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67","Type":"ContainerDied","Data":"20f9fe0648a37b2bdc0698da559206b4abefab1c0fd124f5089588d63d9211e1"} Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.835272 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.835352 4858 scope.go:117] "RemoveContainer" containerID="9afb21abc17b156fb1141a5311a1806842df605aca844698373f63864ed75bec" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.837973 4858 generic.go:334] "Generic (PLEG): container finished" podID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerID="42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188" exitCode=0 Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.838050 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crt7w" event={"ID":"c8456e28-cc53-4820-8bbf-44e27de1dc9b","Type":"ContainerDied","Data":"42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188"} Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.842872 4858 generic.go:334] "Generic (PLEG): container finished" podID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerID="9dd033e13afafac720728adc96412bc8c9145a7837a885562bb5ccd265331652" exitCode=0 Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.842935 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbfmx" event={"ID":"9e6ac9fc-835d-4763-a6f5-e6923a5ee981","Type":"ContainerDied","Data":"9dd033e13afafac720728adc96412bc8c9145a7837a885562bb5ccd265331652"} Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.847929 4858 generic.go:334] "Generic (PLEG): container finished" podID="9daba85d-2681-4f74-8094-9db79d723cee" containerID="97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4" exitCode=0 Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.848020 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kdlf" event={"ID":"9daba85d-2681-4f74-8094-9db79d723cee","Type":"ContainerDied","Data":"97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4"} Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.851634 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" event={"ID":"6b144d61-bc46-4b54-9bd4-7b2127c7ffc6","Type":"ContainerDied","Data":"7c3480dd635356c36be92ece9597e21b8181a6c9a18d88da279cd35ff99876ff"} Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.851708 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.862698 4858 scope.go:117] "RemoveContainer" containerID="c3df189adb4d3b6bd5e784894ec6e2d7aef021b7943893f069781970a606f06d" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.887937 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr"] Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.890273 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.890403 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.896482 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c569f8fcc-6pzmr"] Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.912957 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7"] Mar 20 09:01:37 crc kubenswrapper[4858]: I0320 09:01:37.915431 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77f575b587-swmq7"] Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.080203 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" path="/var/lib/kubelet/pods/6b144d61-bc46-4b54-9bd4-7b2127c7ffc6/volumes" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.080905 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" path="/var/lib/kubelet/pods/e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67/volumes" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.118457 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-589665f995-b84bg"] Mar 20 09:01:38 crc kubenswrapper[4858]: E0320 09:01:38.118893 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" containerName="route-controller-manager" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.118918 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" containerName="route-controller-manager" Mar 20 09:01:38 crc kubenswrapper[4858]: E0320 09:01:38.118939 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fe10ec-a162-4c93-b2d3-1a80745e7fcc" containerName="oc" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.118946 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fe10ec-a162-4c93-b2d3-1a80745e7fcc" containerName="oc" Mar 20 09:01:38 crc kubenswrapper[4858]: E0320 09:01:38.118961 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" containerName="controller-manager" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.118969 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" containerName="controller-manager" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.119078 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1e7c7d3-92e0-4e61-a8f6-28dd00f1ac67" containerName="controller-manager" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.119094 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="74fe10ec-a162-4c93-b2d3-1a80745e7fcc" containerName="oc" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.119109 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b144d61-bc46-4b54-9bd4-7b2127c7ffc6" containerName="route-controller-manager" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.119709 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.123392 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh"] Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.123882 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.124415 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.124543 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.124656 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.124808 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.126306 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.126416 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.133367 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.133512 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.133736 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.133781 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.134078 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.134113 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.136902 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.137184 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh"] Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.138962 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589665f995-b84bg"] Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198259 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-config\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198360 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-serving-cert\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198380 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jczdv\" (UniqueName: \"kubernetes.io/projected/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-kube-api-access-jczdv\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198395 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-client-ca\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198442 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2csn\" (UniqueName: \"kubernetes.io/projected/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-kube-api-access-x2csn\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198461 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-config\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198486 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-proxy-ca-bundles\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198543 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-client-ca\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.198577 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-serving-cert\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.300522 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-client-ca\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.301169 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-serving-cert\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.301225 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-config\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.301260 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-serving-cert\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.301290 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jczdv\" (UniqueName: \"kubernetes.io/projected/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-kube-api-access-jczdv\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.302489 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-config\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.302510 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-client-ca\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.303286 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-client-ca\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.303372 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2csn\" (UniqueName: \"kubernetes.io/projected/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-kube-api-access-x2csn\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.303399 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-config\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.307867 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-config\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.313875 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-client-ca\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.319442 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-proxy-ca-bundles\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.320491 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-serving-cert\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.320599 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-serving-cert\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.324351 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-proxy-ca-bundles\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.326789 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jczdv\" (UniqueName: \"kubernetes.io/projected/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-kube-api-access-jczdv\") pod \"route-controller-manager-598f4fdcc-kl4wh\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.328234 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2csn\" (UniqueName: \"kubernetes.io/projected/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-kube-api-access-x2csn\") pod \"controller-manager-589665f995-b84bg\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.451046 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.459184 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:38 crc kubenswrapper[4858]: I0320 09:01:38.972723 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh"] Mar 20 09:01:38 crc kubenswrapper[4858]: W0320 09:01:38.980401 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a74a3e2_a112_497e_be08_1c8dbb0b2a43.slice/crio-624426a138a911b8ade7672f2a979d24367706666487ef69f345b36ec5a25b4a WatchSource:0}: Error finding container 624426a138a911b8ade7672f2a979d24367706666487ef69f345b36ec5a25b4a: Status 404 returned error can't find the container with id 624426a138a911b8ade7672f2a979d24367706666487ef69f345b36ec5a25b4a Mar 20 09:01:39 crc kubenswrapper[4858]: I0320 09:01:39.023277 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589665f995-b84bg"] Mar 20 09:01:39 crc kubenswrapper[4858]: W0320 09:01:39.027854 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6727c06c_0d58_4feb_a0d7_56a9d0812c7f.slice/crio-f1277a2d6715c8a9fd7c5c2fe2307443ad8117757c88e8c2bd61d7c135f6e281 WatchSource:0}: Error finding container f1277a2d6715c8a9fd7c5c2fe2307443ad8117757c88e8c2bd61d7c135f6e281: Status 404 returned error can't find the container with id f1277a2d6715c8a9fd7c5c2fe2307443ad8117757c88e8c2bd61d7c135f6e281 Mar 20 09:01:39 crc kubenswrapper[4858]: I0320 09:01:39.869297 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" event={"ID":"1a74a3e2-a112-497e-be08-1c8dbb0b2a43","Type":"ContainerStarted","Data":"624426a138a911b8ade7672f2a979d24367706666487ef69f345b36ec5a25b4a"} Mar 20 09:01:39 crc kubenswrapper[4858]: I0320 09:01:39.871201 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" event={"ID":"6727c06c-0d58-4feb-a0d7-56a9d0812c7f","Type":"ContainerStarted","Data":"f1277a2d6715c8a9fd7c5c2fe2307443ad8117757c88e8c2bd61d7c135f6e281"} Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.885180 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbfmx" event={"ID":"9e6ac9fc-835d-4763-a6f5-e6923a5ee981","Type":"ContainerStarted","Data":"288a6a90769ce76c5cf08631afab7f75c5b583ba545374842448f3e542b3398d"} Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.886762 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" event={"ID":"6727c06c-0d58-4feb-a0d7-56a9d0812c7f","Type":"ContainerStarted","Data":"e00c8b6c95407ca00e8215f95666cbc1e6d75376e760f4f0f6ec8b90664af136"} Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.887584 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.889581 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" event={"ID":"1a74a3e2-a112-497e-be08-1c8dbb0b2a43","Type":"ContainerStarted","Data":"da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33"} Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.889843 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.894624 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.895163 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.916650 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tbfmx" podStartSLOduration=4.8911646619999996 podStartE2EDuration="1m1.916633172s" podCreationTimestamp="2026-03-20 09:00:40 +0000 UTC" firstStartedPulling="2026-03-20 09:00:43.230139718 +0000 UTC m=+224.550557915" lastFinishedPulling="2026-03-20 09:01:40.255608218 +0000 UTC m=+281.576026425" observedRunningTime="2026-03-20 09:01:41.914583433 +0000 UTC m=+283.235001650" watchObservedRunningTime="2026-03-20 09:01:41.916633172 +0000 UTC m=+283.237051369" Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.955277 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" podStartSLOduration=5.9552532540000005 podStartE2EDuration="5.955253254s" podCreationTimestamp="2026-03-20 09:01:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:41.953866461 +0000 UTC m=+283.274284668" watchObservedRunningTime="2026-03-20 09:01:41.955253254 +0000 UTC m=+283.275671471" Mar 20 09:01:41 crc kubenswrapper[4858]: I0320 09:01:41.957616 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" podStartSLOduration=5.957607182 podStartE2EDuration="5.957607182s" podCreationTimestamp="2026-03-20 09:01:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:01:41.936691822 +0000 UTC m=+283.257110039" watchObservedRunningTime="2026-03-20 09:01:41.957607182 +0000 UTC m=+283.278025389" Mar 20 09:01:46 crc kubenswrapper[4858]: I0320 09:01:46.931055 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kdlf" event={"ID":"9daba85d-2681-4f74-8094-9db79d723cee","Type":"ContainerStarted","Data":"8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8"} Mar 20 09:01:46 crc kubenswrapper[4858]: I0320 09:01:46.958156 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6kdlf" podStartSLOduration=3.353988834 podStartE2EDuration="1m9.958135683s" podCreationTimestamp="2026-03-20 09:00:37 +0000 UTC" firstStartedPulling="2026-03-20 09:00:40.007341875 +0000 UTC m=+221.327760072" lastFinishedPulling="2026-03-20 09:01:46.611488724 +0000 UTC m=+287.931906921" observedRunningTime="2026-03-20 09:01:46.952074696 +0000 UTC m=+288.272492893" watchObservedRunningTime="2026-03-20 09:01:46.958135683 +0000 UTC m=+288.278553880" Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.553427 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.553513 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.940139 4858 generic.go:334] "Generic (PLEG): container finished" podID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerID="8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89" exitCode=0 Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.940219 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dv2r" event={"ID":"e78c3dad-ee9d-4901-8c08-2db4bd2070cd","Type":"ContainerDied","Data":"8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89"} Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.944142 4858 generic.go:334] "Generic (PLEG): container finished" podID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerID="1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0" exitCode=0 Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.944220 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swvjn" event={"ID":"f98a0de8-b0a6-4c33-83b9-831c88485e50","Type":"ContainerDied","Data":"1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0"} Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.948175 4858 generic.go:334] "Generic (PLEG): container finished" podID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerID="33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13" exitCode=0 Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.948259 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbnzw" event={"ID":"508d2d5b-0a75-4130-a396-9253b685e2cd","Type":"ContainerDied","Data":"33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13"} Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.951519 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crt7w" event={"ID":"c8456e28-cc53-4820-8bbf-44e27de1dc9b","Type":"ContainerStarted","Data":"6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6"} Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.953547 4858 generic.go:334] "Generic (PLEG): container finished" podID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerID="c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d" exitCode=0 Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.953606 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwft" event={"ID":"1b56366e-866a-4139-9b65-3228c5f92d4a","Type":"ContainerDied","Data":"c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d"} Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.958787 4858 generic.go:334] "Generic (PLEG): container finished" podID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerID="57e0f712686bc4cfcc99e302674f5dc390ad37ccb685bfb501aa7f581a45bae0" exitCode=0 Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.958903 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mw84k" event={"ID":"57d2d9c4-3ee7-41f7-af06-18c775cb10c4","Type":"ContainerDied","Data":"57e0f712686bc4cfcc99e302674f5dc390ad37ccb685bfb501aa7f581a45bae0"} Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.979043 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:01:47 crc kubenswrapper[4858]: I0320 09:01:47.979461 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:01:48 crc kubenswrapper[4858]: I0320 09:01:48.066839 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-crt7w" podStartSLOduration=4.379401936 podStartE2EDuration="1m11.066813339s" podCreationTimestamp="2026-03-20 09:00:37 +0000 UTC" firstStartedPulling="2026-03-20 09:00:39.946292243 +0000 UTC m=+221.266710440" lastFinishedPulling="2026-03-20 09:01:46.633703646 +0000 UTC m=+287.954121843" observedRunningTime="2026-03-20 09:01:48.060157587 +0000 UTC m=+289.380575794" watchObservedRunningTime="2026-03-20 09:01:48.066813339 +0000 UTC m=+289.387231546" Mar 20 09:01:48 crc kubenswrapper[4858]: I0320 09:01:48.680242 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6kdlf" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:48 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:48 crc kubenswrapper[4858]: > Mar 20 09:01:48 crc kubenswrapper[4858]: I0320 09:01:48.968265 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dv2r" event={"ID":"e78c3dad-ee9d-4901-8c08-2db4bd2070cd","Type":"ContainerStarted","Data":"b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6"} Mar 20 09:01:48 crc kubenswrapper[4858]: I0320 09:01:48.974404 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swvjn" event={"ID":"f98a0de8-b0a6-4c33-83b9-831c88485e50","Type":"ContainerStarted","Data":"e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31"} Mar 20 09:01:48 crc kubenswrapper[4858]: I0320 09:01:48.976870 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbnzw" event={"ID":"508d2d5b-0a75-4130-a396-9253b685e2cd","Type":"ContainerStarted","Data":"11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392"} Mar 20 09:01:48 crc kubenswrapper[4858]: I0320 09:01:48.979157 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwft" event={"ID":"1b56366e-866a-4139-9b65-3228c5f92d4a","Type":"ContainerStarted","Data":"cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c"} Mar 20 09:01:48 crc kubenswrapper[4858]: I0320 09:01:48.981386 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mw84k" event={"ID":"57d2d9c4-3ee7-41f7-af06-18c775cb10c4","Type":"ContainerStarted","Data":"fbc3c9abd53fedccf29bfdc46ed6c5baffc813a0c2a07eca1ecf6f4b4cab2ffd"} Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.014021 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2dv2r" podStartSLOduration=4.516142093 podStartE2EDuration="1m13.013996094s" podCreationTimestamp="2026-03-20 09:00:36 +0000 UTC" firstStartedPulling="2026-03-20 09:00:39.895155378 +0000 UTC m=+221.215573575" lastFinishedPulling="2026-03-20 09:01:48.393009379 +0000 UTC m=+289.713427576" observedRunningTime="2026-03-20 09:01:48.995910803 +0000 UTC m=+290.316329000" watchObservedRunningTime="2026-03-20 09:01:49.013996094 +0000 UTC m=+290.334414291" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.046708 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mw84k" podStartSLOduration=3.556220029 podStartE2EDuration="1m10.046684532s" podCreationTimestamp="2026-03-20 09:00:39 +0000 UTC" firstStartedPulling="2026-03-20 09:00:42.180053739 +0000 UTC m=+223.500471936" lastFinishedPulling="2026-03-20 09:01:48.670518242 +0000 UTC m=+289.990936439" observedRunningTime="2026-03-20 09:01:49.022340108 +0000 UTC m=+290.342758305" watchObservedRunningTime="2026-03-20 09:01:49.046684532 +0000 UTC m=+290.367102729" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.049335 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-swvjn" podStartSLOduration=4.789886526 podStartE2EDuration="1m9.049307246s" podCreationTimestamp="2026-03-20 09:00:40 +0000 UTC" firstStartedPulling="2026-03-20 09:00:44.25487079 +0000 UTC m=+225.575288987" lastFinishedPulling="2026-03-20 09:01:48.51429151 +0000 UTC m=+289.834709707" observedRunningTime="2026-03-20 09:01:49.0437249 +0000 UTC m=+290.364143107" watchObservedRunningTime="2026-03-20 09:01:49.049307246 +0000 UTC m=+290.369725443" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.052221 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-crt7w" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:49 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:49 crc kubenswrapper[4858]: > Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.071645 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4dwft" podStartSLOduration=3.498496054 podStartE2EDuration="1m12.071620221s" podCreationTimestamp="2026-03-20 09:00:37 +0000 UTC" firstStartedPulling="2026-03-20 09:00:40.012062143 +0000 UTC m=+221.332480340" lastFinishedPulling="2026-03-20 09:01:48.58518631 +0000 UTC m=+289.905604507" observedRunningTime="2026-03-20 09:01:49.070831891 +0000 UTC m=+290.391250088" watchObservedRunningTime="2026-03-20 09:01:49.071620221 +0000 UTC m=+290.392038438" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.090631 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pbnzw" podStartSLOduration=2.833240187 podStartE2EDuration="1m10.090610494s" podCreationTimestamp="2026-03-20 09:00:39 +0000 UTC" firstStartedPulling="2026-03-20 09:00:41.070247887 +0000 UTC m=+222.390666084" lastFinishedPulling="2026-03-20 09:01:48.327618184 +0000 UTC m=+289.648036391" observedRunningTime="2026-03-20 09:01:49.088180234 +0000 UTC m=+290.408598451" watchObservedRunningTime="2026-03-20 09:01:49.090610494 +0000 UTC m=+290.411028691" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.598703 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.598776 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.939369 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:01:49 crc kubenswrapper[4858]: I0320 09:01:49.939445 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:01:50 crc kubenswrapper[4858]: I0320 09:01:50.568287 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:01:50 crc kubenswrapper[4858]: I0320 09:01:50.569237 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:01:50 crc kubenswrapper[4858]: I0320 09:01:50.651382 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-pbnzw" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:50 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:50 crc kubenswrapper[4858]: > Mar 20 09:01:50 crc kubenswrapper[4858]: I0320 09:01:50.862813 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:01:50 crc kubenswrapper[4858]: I0320 09:01:50.862895 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:01:50 crc kubenswrapper[4858]: I0320 09:01:50.984360 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-mw84k" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:50 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:50 crc kubenswrapper[4858]: > Mar 20 09:01:51 crc kubenswrapper[4858]: I0320 09:01:51.622447 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-swvjn" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:51 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:51 crc kubenswrapper[4858]: > Mar 20 09:01:51 crc kubenswrapper[4858]: I0320 09:01:51.909259 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tbfmx" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="registry-server" probeResult="failure" output=< Mar 20 09:01:51 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:01:51 crc kubenswrapper[4858]: > Mar 20 09:01:56 crc kubenswrapper[4858]: I0320 09:01:56.349716 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589665f995-b84bg"] Mar 20 09:01:56 crc kubenswrapper[4858]: I0320 09:01:56.350043 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" podUID="6727c06c-0d58-4feb-a0d7-56a9d0812c7f" containerName="controller-manager" containerID="cri-o://e00c8b6c95407ca00e8215f95666cbc1e6d75376e760f4f0f6ec8b90664af136" gracePeriod=30 Mar 20 09:01:56 crc kubenswrapper[4858]: I0320 09:01:56.465524 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh"] Mar 20 09:01:56 crc kubenswrapper[4858]: I0320 09:01:56.465818 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" podUID="1a74a3e2-a112-497e-be08-1c8dbb0b2a43" containerName="route-controller-manager" containerID="cri-o://da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33" gracePeriod=30 Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.015700 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.047764 4858 generic.go:334] "Generic (PLEG): container finished" podID="1a74a3e2-a112-497e-be08-1c8dbb0b2a43" containerID="da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33" exitCode=0 Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.047853 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.047930 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" event={"ID":"1a74a3e2-a112-497e-be08-1c8dbb0b2a43","Type":"ContainerDied","Data":"da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33"} Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.047974 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh" event={"ID":"1a74a3e2-a112-497e-be08-1c8dbb0b2a43","Type":"ContainerDied","Data":"624426a138a911b8ade7672f2a979d24367706666487ef69f345b36ec5a25b4a"} Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.047998 4858 scope.go:117] "RemoveContainer" containerID="da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.050708 4858 generic.go:334] "Generic (PLEG): container finished" podID="6727c06c-0d58-4feb-a0d7-56a9d0812c7f" containerID="e00c8b6c95407ca00e8215f95666cbc1e6d75376e760f4f0f6ec8b90664af136" exitCode=0 Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.050794 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" event={"ID":"6727c06c-0d58-4feb-a0d7-56a9d0812c7f","Type":"ContainerDied","Data":"e00c8b6c95407ca00e8215f95666cbc1e6d75376e760f4f0f6ec8b90664af136"} Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.073221 4858 scope.go:117] "RemoveContainer" containerID="da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33" Mar 20 09:01:57 crc kubenswrapper[4858]: E0320 09:01:57.074238 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33\": container with ID starting with da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33 not found: ID does not exist" containerID="da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.074271 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33"} err="failed to get container status \"da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33\": rpc error: code = NotFound desc = could not find container \"da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33\": container with ID starting with da76346befbdb71c38173ec29a20933db182bcac6023fd3aba5427cc3618eb33 not found: ID does not exist" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.127350 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-config\") pod \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.127414 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-client-ca\") pod \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.127450 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-serving-cert\") pod \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.127528 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jczdv\" (UniqueName: \"kubernetes.io/projected/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-kube-api-access-jczdv\") pod \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\" (UID: \"1a74a3e2-a112-497e-be08-1c8dbb0b2a43\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.128955 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-client-ca" (OuterVolumeSpecName: "client-ca") pod "1a74a3e2-a112-497e-be08-1c8dbb0b2a43" (UID: "1a74a3e2-a112-497e-be08-1c8dbb0b2a43"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.129167 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-config" (OuterVolumeSpecName: "config") pod "1a74a3e2-a112-497e-be08-1c8dbb0b2a43" (UID: "1a74a3e2-a112-497e-be08-1c8dbb0b2a43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.134487 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-kube-api-access-jczdv" (OuterVolumeSpecName: "kube-api-access-jczdv") pod "1a74a3e2-a112-497e-be08-1c8dbb0b2a43" (UID: "1a74a3e2-a112-497e-be08-1c8dbb0b2a43"). InnerVolumeSpecName "kube-api-access-jczdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.134594 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1a74a3e2-a112-497e-be08-1c8dbb0b2a43" (UID: "1a74a3e2-a112-497e-be08-1c8dbb0b2a43"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.230080 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.230160 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.230188 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.230215 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jczdv\" (UniqueName: \"kubernetes.io/projected/1a74a3e2-a112-497e-be08-1c8dbb0b2a43-kube-api-access-jczdv\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.276710 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.276785 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.327264 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.376825 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh"] Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.383572 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-598f4fdcc-kl4wh"] Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.444976 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.538140 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-config\") pod \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.538241 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-proxy-ca-bundles\") pod \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.538292 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-client-ca\") pod \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.538368 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-serving-cert\") pod \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.538412 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2csn\" (UniqueName: \"kubernetes.io/projected/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-kube-api-access-x2csn\") pod \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\" (UID: \"6727c06c-0d58-4feb-a0d7-56a9d0812c7f\") " Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.539210 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-config" (OuterVolumeSpecName: "config") pod "6727c06c-0d58-4feb-a0d7-56a9d0812c7f" (UID: "6727c06c-0d58-4feb-a0d7-56a9d0812c7f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.539243 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6727c06c-0d58-4feb-a0d7-56a9d0812c7f" (UID: "6727c06c-0d58-4feb-a0d7-56a9d0812c7f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.539617 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-client-ca" (OuterVolumeSpecName: "client-ca") pod "6727c06c-0d58-4feb-a0d7-56a9d0812c7f" (UID: "6727c06c-0d58-4feb-a0d7-56a9d0812c7f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.543004 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-kube-api-access-x2csn" (OuterVolumeSpecName: "kube-api-access-x2csn") pod "6727c06c-0d58-4feb-a0d7-56a9d0812c7f" (UID: "6727c06c-0d58-4feb-a0d7-56a9d0812c7f"). InnerVolumeSpecName "kube-api-access-x2csn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.543410 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6727c06c-0d58-4feb-a0d7-56a9d0812c7f" (UID: "6727c06c-0d58-4feb-a0d7-56a9d0812c7f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.590253 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.632102 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.640391 4858 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-client-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.640413 4858 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.640423 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2csn\" (UniqueName: \"kubernetes.io/projected/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-kube-api-access-x2csn\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.640433 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.640442 4858 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6727c06c-0d58-4feb-a0d7-56a9d0812c7f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.883904 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.883984 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:01:57 crc kubenswrapper[4858]: I0320 09:01:57.927712 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.030796 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.075237 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.092850 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a74a3e2-a112-497e-be08-1c8dbb0b2a43" path="/var/lib/kubelet/pods/1a74a3e2-a112-497e-be08-1c8dbb0b2a43/volumes" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.093648 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589665f995-b84bg" event={"ID":"6727c06c-0d58-4feb-a0d7-56a9d0812c7f","Type":"ContainerDied","Data":"f1277a2d6715c8a9fd7c5c2fe2307443ad8117757c88e8c2bd61d7c135f6e281"} Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.093737 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.093781 4858 scope.go:117] "RemoveContainer" containerID="e00c8b6c95407ca00e8215f95666cbc1e6d75376e760f4f0f6ec8b90664af136" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.130374 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd"] Mar 20 09:01:58 crc kubenswrapper[4858]: E0320 09:01:58.130840 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a74a3e2-a112-497e-be08-1c8dbb0b2a43" containerName="route-controller-manager" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.130863 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a74a3e2-a112-497e-be08-1c8dbb0b2a43" containerName="route-controller-manager" Mar 20 09:01:58 crc kubenswrapper[4858]: E0320 09:01:58.130873 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6727c06c-0d58-4feb-a0d7-56a9d0812c7f" containerName="controller-manager" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.130882 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6727c06c-0d58-4feb-a0d7-56a9d0812c7f" containerName="controller-manager" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.131034 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="6727c06c-0d58-4feb-a0d7-56a9d0812c7f" containerName="controller-manager" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.131057 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a74a3e2-a112-497e-be08-1c8dbb0b2a43" containerName="route-controller-manager" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.131805 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.134845 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.136877 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.137305 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75494f854c-q7xx7"] Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.138278 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.138648 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.139158 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.139406 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.139583 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.142212 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.142659 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.142892 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.143062 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.143348 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.144809 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.148225 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589665f995-b84bg"] Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.152952 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd"] Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.155001 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.156188 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.157494 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-589665f995-b84bg"] Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.162865 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75494f854c-q7xx7"] Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.168985 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.248645 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-config\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.248728 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svj6j\" (UniqueName: \"kubernetes.io/projected/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-kube-api-access-svj6j\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.248821 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/446be6a7-dd0d-435b-97a3-58a01471c992-serving-cert\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.248852 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-client-ca\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.248893 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-serving-cert\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.248921 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/446be6a7-dd0d-435b-97a3-58a01471c992-client-ca\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.248954 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/446be6a7-dd0d-435b-97a3-58a01471c992-config\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.249008 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8vmk\" (UniqueName: \"kubernetes.io/projected/446be6a7-dd0d-435b-97a3-58a01471c992-kube-api-access-k8vmk\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.249058 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-proxy-ca-bundles\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.350563 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-config\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.350632 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svj6j\" (UniqueName: \"kubernetes.io/projected/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-kube-api-access-svj6j\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.350678 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/446be6a7-dd0d-435b-97a3-58a01471c992-serving-cert\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.350703 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-client-ca\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.350731 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-serving-cert\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.351976 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-client-ca\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.351561 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/446be6a7-dd0d-435b-97a3-58a01471c992-client-ca\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.352561 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/446be6a7-dd0d-435b-97a3-58a01471c992-client-ca\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.352442 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-config\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.352626 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/446be6a7-dd0d-435b-97a3-58a01471c992-config\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.353021 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8vmk\" (UniqueName: \"kubernetes.io/projected/446be6a7-dd0d-435b-97a3-58a01471c992-kube-api-access-k8vmk\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.353097 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-proxy-ca-bundles\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.354030 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-proxy-ca-bundles\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.355717 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-serving-cert\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.356528 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/446be6a7-dd0d-435b-97a3-58a01471c992-config\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.357378 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/446be6a7-dd0d-435b-97a3-58a01471c992-serving-cert\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.371086 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8vmk\" (UniqueName: \"kubernetes.io/projected/446be6a7-dd0d-435b-97a3-58a01471c992-kube-api-access-k8vmk\") pod \"route-controller-manager-6fdf869f89-9q6sd\" (UID: \"446be6a7-dd0d-435b-97a3-58a01471c992\") " pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.373457 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svj6j\" (UniqueName: \"kubernetes.io/projected/47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a-kube-api-access-svj6j\") pod \"controller-manager-75494f854c-q7xx7\" (UID: \"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a\") " pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.458250 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.465464 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.904704 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75494f854c-q7xx7"] Mar 20 09:01:58 crc kubenswrapper[4858]: W0320 09:01:58.912599 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47d45dd9_d0d5_4e8b_97bb_12b88a6dd18a.slice/crio-f2e7e490542ba9e39f3eb47346d5087ac9e38266380722134a4b199834ae710d WatchSource:0}: Error finding container f2e7e490542ba9e39f3eb47346d5087ac9e38266380722134a4b199834ae710d: Status 404 returned error can't find the container with id f2e7e490542ba9e39f3eb47346d5087ac9e38266380722134a4b199834ae710d Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.963232 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd"] Mar 20 09:01:58 crc kubenswrapper[4858]: I0320 09:01:58.968353 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4dwft"] Mar 20 09:01:59 crc kubenswrapper[4858]: I0320 09:01:59.100168 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" event={"ID":"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a","Type":"ContainerStarted","Data":"f2e7e490542ba9e39f3eb47346d5087ac9e38266380722134a4b199834ae710d"} Mar 20 09:01:59 crc kubenswrapper[4858]: I0320 09:01:59.102086 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2c66r"] Mar 20 09:01:59 crc kubenswrapper[4858]: I0320 09:01:59.105397 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" event={"ID":"446be6a7-dd0d-435b-97a3-58a01471c992","Type":"ContainerStarted","Data":"9e589b41bac264244175b195e0bf0020e08f55455cef5b9be9ec28cbe75e676e"} Mar 20 09:01:59 crc kubenswrapper[4858]: I0320 09:01:59.652030 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:01:59 crc kubenswrapper[4858]: I0320 09:01:59.696861 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.009955 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.057243 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.082393 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6727c06c-0d58-4feb-a0d7-56a9d0812c7f" path="/var/lib/kubelet/pods/6727c06c-0d58-4feb-a0d7-56a9d0812c7f/volumes" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.114571 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" event={"ID":"47d45dd9-d0d5-4e8b-97bb-12b88a6dd18a","Type":"ContainerStarted","Data":"ebe0175eb02a6ca317afcdd689d72c6ef5249393257d11d4781ff5cc026689ce"} Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.115892 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.131847 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" event={"ID":"446be6a7-dd0d-435b-97a3-58a01471c992","Type":"ContainerStarted","Data":"2501883004e12fa7f7b3fbad1e5fcb11fa9935846e3f7799c8cf1c0e126fef42"} Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.132406 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4dwft" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="registry-server" containerID="cri-o://cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c" gracePeriod=2 Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.133696 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.135840 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.143416 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.156936 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566622-l8kbk"] Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.161778 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-l8kbk"] Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.161935 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.165619 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.165881 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.166055 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.211355 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75494f854c-q7xx7" podStartSLOduration=4.211310694 podStartE2EDuration="4.211310694s" podCreationTimestamp="2026-03-20 09:01:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:02:00.204095152 +0000 UTC m=+301.524513359" watchObservedRunningTime="2026-03-20 09:02:00.211310694 +0000 UTC m=+301.531728901" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.255610 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6fdf869f89-9q6sd" podStartSLOduration=4.255577465 podStartE2EDuration="4.255577465s" podCreationTimestamp="2026-03-20 09:01:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:02:00.228884166 +0000 UTC m=+301.549302383" watchObservedRunningTime="2026-03-20 09:02:00.255577465 +0000 UTC m=+301.575995662" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.284585 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x66z\" (UniqueName: \"kubernetes.io/projected/803f2926-2469-4a09-85ba-a1c3e4548168-kube-api-access-4x66z\") pod \"auto-csr-approver-29566622-l8kbk\" (UID: \"803f2926-2469-4a09-85ba-a1c3e4548168\") " pod="openshift-infra/auto-csr-approver-29566622-l8kbk" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.365054 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-crt7w"] Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.365573 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-crt7w" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="registry-server" containerID="cri-o://6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6" gracePeriod=2 Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.385884 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x66z\" (UniqueName: \"kubernetes.io/projected/803f2926-2469-4a09-85ba-a1c3e4548168-kube-api-access-4x66z\") pod \"auto-csr-approver-29566622-l8kbk\" (UID: \"803f2926-2469-4a09-85ba-a1c3e4548168\") " pod="openshift-infra/auto-csr-approver-29566622-l8kbk" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.407474 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x66z\" (UniqueName: \"kubernetes.io/projected/803f2926-2469-4a09-85ba-a1c3e4548168-kube-api-access-4x66z\") pod \"auto-csr-approver-29566622-l8kbk\" (UID: \"803f2926-2469-4a09-85ba-a1c3e4548168\") " pod="openshift-infra/auto-csr-approver-29566622-l8kbk" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.519745 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.629266 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.694683 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.847500 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.895761 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.912346 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.960671 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.982246 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-l8kbk"] Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.994728 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g575q\" (UniqueName: \"kubernetes.io/projected/1b56366e-866a-4139-9b65-3228c5f92d4a-kube-api-access-g575q\") pod \"1b56366e-866a-4139-9b65-3228c5f92d4a\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.994835 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-utilities\") pod \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.995078 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-utilities\") pod \"1b56366e-866a-4139-9b65-3228c5f92d4a\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.995113 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-catalog-content\") pod \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.995143 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9gqc\" (UniqueName: \"kubernetes.io/projected/c8456e28-cc53-4820-8bbf-44e27de1dc9b-kube-api-access-n9gqc\") pod \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\" (UID: \"c8456e28-cc53-4820-8bbf-44e27de1dc9b\") " Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.995219 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-catalog-content\") pod \"1b56366e-866a-4139-9b65-3228c5f92d4a\" (UID: \"1b56366e-866a-4139-9b65-3228c5f92d4a\") " Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.996101 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-utilities" (OuterVolumeSpecName: "utilities") pod "1b56366e-866a-4139-9b65-3228c5f92d4a" (UID: "1b56366e-866a-4139-9b65-3228c5f92d4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:00 crc kubenswrapper[4858]: I0320 09:02:00.996391 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-utilities" (OuterVolumeSpecName: "utilities") pod "c8456e28-cc53-4820-8bbf-44e27de1dc9b" (UID: "c8456e28-cc53-4820-8bbf-44e27de1dc9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.002586 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8456e28-cc53-4820-8bbf-44e27de1dc9b-kube-api-access-n9gqc" (OuterVolumeSpecName: "kube-api-access-n9gqc") pod "c8456e28-cc53-4820-8bbf-44e27de1dc9b" (UID: "c8456e28-cc53-4820-8bbf-44e27de1dc9b"). InnerVolumeSpecName "kube-api-access-n9gqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.003887 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b56366e-866a-4139-9b65-3228c5f92d4a-kube-api-access-g575q" (OuterVolumeSpecName: "kube-api-access-g575q") pod "1b56366e-866a-4139-9b65-3228c5f92d4a" (UID: "1b56366e-866a-4139-9b65-3228c5f92d4a"). InnerVolumeSpecName "kube-api-access-g575q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.052144 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8456e28-cc53-4820-8bbf-44e27de1dc9b" (UID: "c8456e28-cc53-4820-8bbf-44e27de1dc9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.097533 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g575q\" (UniqueName: \"kubernetes.io/projected/1b56366e-866a-4139-9b65-3228c5f92d4a-kube-api-access-g575q\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.097607 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.097625 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.097638 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8456e28-cc53-4820-8bbf-44e27de1dc9b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.097654 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9gqc\" (UniqueName: \"kubernetes.io/projected/c8456e28-cc53-4820-8bbf-44e27de1dc9b-kube-api-access-n9gqc\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.142250 4858 generic.go:334] "Generic (PLEG): container finished" podID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerID="cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c" exitCode=0 Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.142373 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwft" event={"ID":"1b56366e-866a-4139-9b65-3228c5f92d4a","Type":"ContainerDied","Data":"cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c"} Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.142494 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4dwft" event={"ID":"1b56366e-866a-4139-9b65-3228c5f92d4a","Type":"ContainerDied","Data":"56189411e3d4ebd6d6333111a6e34c5f42a4d5c36779effda3b59e8fdf1380a5"} Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.142419 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4dwft" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.142555 4858 scope.go:117] "RemoveContainer" containerID="cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.145457 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" event={"ID":"803f2926-2469-4a09-85ba-a1c3e4548168","Type":"ContainerStarted","Data":"cb196fba725d47d9961e9c9cb91387ac9b319c178007e8b2a8523199f51de20d"} Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.149811 4858 generic.go:334] "Generic (PLEG): container finished" podID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerID="6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6" exitCode=0 Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.150166 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crt7w" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.150185 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crt7w" event={"ID":"c8456e28-cc53-4820-8bbf-44e27de1dc9b","Type":"ContainerDied","Data":"6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6"} Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.150304 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crt7w" event={"ID":"c8456e28-cc53-4820-8bbf-44e27de1dc9b","Type":"ContainerDied","Data":"45d9e21713de5685e5de8d219f56ded96a618ff8782d0482c8a18da0de683f6c"} Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.163790 4858 scope.go:117] "RemoveContainer" containerID="c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.183141 4858 scope.go:117] "RemoveContainer" containerID="0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.184540 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-crt7w"] Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.187695 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-crt7w"] Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.210992 4858 scope.go:117] "RemoveContainer" containerID="cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c" Mar 20 09:02:01 crc kubenswrapper[4858]: E0320 09:02:01.211676 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c\": container with ID starting with cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c not found: ID does not exist" containerID="cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.211725 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c"} err="failed to get container status \"cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c\": rpc error: code = NotFound desc = could not find container \"cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c\": container with ID starting with cf196c885ceb427cbca4bde7295675841769e1abbb89a69e8616d6ee22cb404c not found: ID does not exist" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.211760 4858 scope.go:117] "RemoveContainer" containerID="c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d" Mar 20 09:02:01 crc kubenswrapper[4858]: E0320 09:02:01.212023 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d\": container with ID starting with c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d not found: ID does not exist" containerID="c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.212052 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d"} err="failed to get container status \"c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d\": rpc error: code = NotFound desc = could not find container \"c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d\": container with ID starting with c55dc6291b82883f46cd26f8f2a5c641a23b7f85a70d676cafa884764b70a74d not found: ID does not exist" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.212066 4858 scope.go:117] "RemoveContainer" containerID="0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b" Mar 20 09:02:01 crc kubenswrapper[4858]: E0320 09:02:01.212280 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b\": container with ID starting with 0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b not found: ID does not exist" containerID="0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.212342 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b"} err="failed to get container status \"0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b\": rpc error: code = NotFound desc = could not find container \"0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b\": container with ID starting with 0397be7b74513c496e503005b1f87de7bf973d30e31d9befd98170a33fd2267b not found: ID does not exist" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.212358 4858 scope.go:117] "RemoveContainer" containerID="6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.227574 4858 scope.go:117] "RemoveContainer" containerID="42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.246628 4858 scope.go:117] "RemoveContainer" containerID="f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.276025 4858 scope.go:117] "RemoveContainer" containerID="6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6" Mar 20 09:02:01 crc kubenswrapper[4858]: E0320 09:02:01.277110 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6\": container with ID starting with 6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6 not found: ID does not exist" containerID="6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.277253 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6"} err="failed to get container status \"6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6\": rpc error: code = NotFound desc = could not find container \"6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6\": container with ID starting with 6b5907a70cd3b97800c544aec7511556fcf9bde0eb84011ba96e8907fded2ec6 not found: ID does not exist" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.277408 4858 scope.go:117] "RemoveContainer" containerID="42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188" Mar 20 09:02:01 crc kubenswrapper[4858]: E0320 09:02:01.278043 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188\": container with ID starting with 42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188 not found: ID does not exist" containerID="42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.278115 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188"} err="failed to get container status \"42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188\": rpc error: code = NotFound desc = could not find container \"42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188\": container with ID starting with 42aa1adfe65d61d7235290842b74a32e3afa4469c97de1c8006daf1554738188 not found: ID does not exist" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.278158 4858 scope.go:117] "RemoveContainer" containerID="f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3" Mar 20 09:02:01 crc kubenswrapper[4858]: E0320 09:02:01.278560 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3\": container with ID starting with f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3 not found: ID does not exist" containerID="f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.278587 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3"} err="failed to get container status \"f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3\": rpc error: code = NotFound desc = could not find container \"f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3\": container with ID starting with f90a38dee38b1da799bca8317779a93a2d7092dbee4f02067e5ffecdd9f39bd3 not found: ID does not exist" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.659666 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b56366e-866a-4139-9b65-3228c5f92d4a" (UID: "1b56366e-866a-4139-9b65-3228c5f92d4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.706672 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b56366e-866a-4139-9b65-3228c5f92d4a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.781603 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4dwft"] Mar 20 09:02:01 crc kubenswrapper[4858]: I0320 09:02:01.791576 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4dwft"] Mar 20 09:02:02 crc kubenswrapper[4858]: I0320 09:02:02.080732 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" path="/var/lib/kubelet/pods/1b56366e-866a-4139-9b65-3228c5f92d4a/volumes" Mar 20 09:02:02 crc kubenswrapper[4858]: I0320 09:02:02.081990 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" path="/var/lib/kubelet/pods/c8456e28-cc53-4820-8bbf-44e27de1dc9b/volumes" Mar 20 09:02:02 crc kubenswrapper[4858]: I0320 09:02:02.765594 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mw84k"] Mar 20 09:02:02 crc kubenswrapper[4858]: I0320 09:02:02.766257 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mw84k" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="registry-server" containerID="cri-o://fbc3c9abd53fedccf29bfdc46ed6c5baffc813a0c2a07eca1ecf6f4b4cab2ffd" gracePeriod=2 Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.176170 4858 generic.go:334] "Generic (PLEG): container finished" podID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerID="fbc3c9abd53fedccf29bfdc46ed6c5baffc813a0c2a07eca1ecf6f4b4cab2ffd" exitCode=0 Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.176233 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mw84k" event={"ID":"57d2d9c4-3ee7-41f7-af06-18c775cb10c4","Type":"ContainerDied","Data":"fbc3c9abd53fedccf29bfdc46ed6c5baffc813a0c2a07eca1ecf6f4b4cab2ffd"} Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.364529 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.542397 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4jf2\" (UniqueName: \"kubernetes.io/projected/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-kube-api-access-p4jf2\") pod \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.542622 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-catalog-content\") pod \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.542655 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-utilities\") pod \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\" (UID: \"57d2d9c4-3ee7-41f7-af06-18c775cb10c4\") " Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.543975 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-utilities" (OuterVolumeSpecName: "utilities") pod "57d2d9c4-3ee7-41f7-af06-18c775cb10c4" (UID: "57d2d9c4-3ee7-41f7-af06-18c775cb10c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.548303 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-kube-api-access-p4jf2" (OuterVolumeSpecName: "kube-api-access-p4jf2") pod "57d2d9c4-3ee7-41f7-af06-18c775cb10c4" (UID: "57d2d9c4-3ee7-41f7-af06-18c775cb10c4"). InnerVolumeSpecName "kube-api-access-p4jf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.574582 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57d2d9c4-3ee7-41f7-af06-18c775cb10c4" (UID: "57d2d9c4-3ee7-41f7-af06-18c775cb10c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.603584 4858 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604000 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="extract-utilities" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604019 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="extract-utilities" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604035 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="extract-utilities" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604043 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="extract-utilities" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604056 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="extract-content" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604064 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="extract-content" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604073 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="extract-content" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604079 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="extract-content" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604089 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="extract-content" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604098 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="extract-content" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604110 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604116 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604129 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="extract-utilities" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604174 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="extract-utilities" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604187 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604194 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.604203 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604211 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604365 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8456e28-cc53-4820-8bbf-44e27de1dc9b" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604386 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604395 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b56366e-866a-4139-9b65-3228c5f92d4a" containerName="registry-server" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.604895 4858 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.605186 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9" gracePeriod=15 Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.605414 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.606100 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859" gracePeriod=15 Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.606402 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc" gracePeriod=15 Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.606378 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0" gracePeriod=15 Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.606397 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5" gracePeriod=15 Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.606805 4858 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607089 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607112 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607121 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607128 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607138 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607145 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607152 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607159 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607167 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607174 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607216 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607222 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607232 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607238 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607247 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607257 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607275 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607285 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 09:02:03 crc kubenswrapper[4858]: E0320 09:02:03.607296 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607303 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607935 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607951 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607959 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607966 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607976 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607985 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.607991 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.608000 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.608220 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.645703 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.645739 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.645751 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4jf2\" (UniqueName: \"kubernetes.io/projected/57d2d9c4-3ee7-41f7-af06-18c775cb10c4-kube-api-access-p4jf2\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.738193 4858 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.738279 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.746910 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.747023 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.747066 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.747086 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.747108 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.747132 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.747158 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.747177 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848264 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848386 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848427 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848477 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848512 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848585 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848577 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848616 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848788 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848835 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848841 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848874 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848897 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848961 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.848989 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:03 crc kubenswrapper[4858]: I0320 09:02:03.849087 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.031094 4858 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.031196 4858 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.184104 4858 generic.go:334] "Generic (PLEG): container finished" podID="803f2926-2469-4a09-85ba-a1c3e4548168" containerID="697e39d85e38f417fa41aff2272b77f57de149bb17df404e92c2553d9fc17a58" exitCode=0 Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.184634 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" event={"ID":"803f2926-2469-4a09-85ba-a1c3e4548168","Type":"ContainerDied","Data":"697e39d85e38f417fa41aff2272b77f57de149bb17df404e92c2553d9fc17a58"} Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.184934 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.188445 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.190421 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.199436 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859" exitCode=0 Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.199482 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc" exitCode=0 Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.199494 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5" exitCode=0 Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.199503 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0" exitCode=2 Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.199550 4858 scope.go:117] "RemoveContainer" containerID="a59c731955e0246515611eb0045fbcf83411f3e3c14689cd5fb58a9103655208" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.203628 4858 generic.go:334] "Generic (PLEG): container finished" podID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" containerID="4ca8c3ddec56c38859968f4560a7a3d304b47b5d70d59813d00b4ba6a528155d" exitCode=0 Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.203701 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b42b85d2-a66d-40be-af75-611eaa9d1a3a","Type":"ContainerDied","Data":"4ca8c3ddec56c38859968f4560a7a3d304b47b5d70d59813d00b4ba6a528155d"} Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.204587 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.205024 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.207678 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mw84k" event={"ID":"57d2d9c4-3ee7-41f7-af06-18c775cb10c4","Type":"ContainerDied","Data":"867facedf8678579b70edeb18020c2d610d9275e9618ff1ff9e6d6ea47fa0786"} Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.207823 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mw84k" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.208770 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.209213 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.209820 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.242518 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.243165 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.243750 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.251435 4858 scope.go:117] "RemoveContainer" containerID="fbc3c9abd53fedccf29bfdc46ed6c5baffc813a0c2a07eca1ecf6f4b4cab2ffd" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.268499 4858 scope.go:117] "RemoveContainer" containerID="57e0f712686bc4cfcc99e302674f5dc390ad37ccb685bfb501aa7f581a45bae0" Mar 20 09:02:04 crc kubenswrapper[4858]: I0320 09:02:04.291418 4858 scope.go:117] "RemoveContainer" containerID="5777b847f61be65de3e998433933800950cc4d41455e48965afca5581c8c4075" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.222506 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.642731 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.643626 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.644000 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.644505 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.648885 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.649705 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.650080 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.650494 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.789345 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x66z\" (UniqueName: \"kubernetes.io/projected/803f2926-2469-4a09-85ba-a1c3e4548168-kube-api-access-4x66z\") pod \"803f2926-2469-4a09-85ba-a1c3e4548168\" (UID: \"803f2926-2469-4a09-85ba-a1c3e4548168\") " Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.790359 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kube-api-access\") pod \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.790539 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kubelet-dir\") pod \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.790746 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-var-lock\") pod \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\" (UID: \"b42b85d2-a66d-40be-af75-611eaa9d1a3a\") " Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.790609 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b42b85d2-a66d-40be-af75-611eaa9d1a3a" (UID: "b42b85d2-a66d-40be-af75-611eaa9d1a3a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.791681 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-var-lock" (OuterVolumeSpecName: "var-lock") pod "b42b85d2-a66d-40be-af75-611eaa9d1a3a" (UID: "b42b85d2-a66d-40be-af75-611eaa9d1a3a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.797653 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/803f2926-2469-4a09-85ba-a1c3e4548168-kube-api-access-4x66z" (OuterVolumeSpecName: "kube-api-access-4x66z") pod "803f2926-2469-4a09-85ba-a1c3e4548168" (UID: "803f2926-2469-4a09-85ba-a1c3e4548168"). InnerVolumeSpecName "kube-api-access-4x66z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.815546 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b42b85d2-a66d-40be-af75-611eaa9d1a3a" (UID: "b42b85d2-a66d-40be-af75-611eaa9d1a3a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.892447 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.892487 4858 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.892495 4858 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b42b85d2-a66d-40be-af75-611eaa9d1a3a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:05 crc kubenswrapper[4858]: I0320 09:02:05.892506 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x66z\" (UniqueName: \"kubernetes.io/projected/803f2926-2469-4a09-85ba-a1c3e4548168-kube-api-access-4x66z\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.235442 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" event={"ID":"803f2926-2469-4a09-85ba-a1c3e4548168","Type":"ContainerDied","Data":"cb196fba725d47d9961e9c9cb91387ac9b319c178007e8b2a8523199f51de20d"} Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.235489 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.235508 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb196fba725d47d9961e9c9cb91387ac9b319c178007e8b2a8523199f51de20d" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.239446 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.240412 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.240467 4858 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9" exitCode=0 Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.240872 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.241131 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.242491 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b42b85d2-a66d-40be-af75-611eaa9d1a3a","Type":"ContainerDied","Data":"48190b4afb2273f4f4f21d5680c1722b840cbc74d8ec5340da4b421203d71503"} Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.242529 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48190b4afb2273f4f4f21d5680c1722b840cbc74d8ec5340da4b421203d71503" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.242556 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.245749 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.246090 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.246461 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.466609 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.468275 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.469430 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.469993 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.470447 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.470775 4858 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603101 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603178 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603269 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603305 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603380 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603458 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603849 4858 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603873 4858 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:06 crc kubenswrapper[4858]: I0320 09:02:06.603886 4858 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.252360 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.253504 4858 scope.go:117] "RemoveContainer" containerID="74ec0eb71b8068c6c5e36e1c696f9314c0396bcf3989050bd9dbb9365a07b859" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.253617 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.271423 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.271856 4858 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.272108 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.272599 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.273760 4858 scope.go:117] "RemoveContainer" containerID="653f5e0c6f0968bb964388c27744d83f2918991a87d1e1eb75c001ac8f3e1efc" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.295886 4858 scope.go:117] "RemoveContainer" containerID="55093dd7960e768725785925ec725669cd67433de9ae3daba579c252e8b5cab5" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.323496 4858 scope.go:117] "RemoveContainer" containerID="fe65a2f0c1f960e11a102907c167ef89cb668363e8264842aee6254bbbc9d2d0" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.343197 4858 scope.go:117] "RemoveContainer" containerID="4aafb8bc4daf95bc8a6f7e10f96998f326369fdbccd6c41bfee4d317fae23ac9" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.363146 4858 scope.go:117] "RemoveContainer" containerID="45a98878d1cbbf3432427a74fddea7a54ab0593e8e1e1da7747569fd5d578cca" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.890710 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.890870 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.890983 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.892413 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:02:07 crc kubenswrapper[4858]: I0320 09:02:07.892561 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd" gracePeriod=600 Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.077945 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.261541 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd" exitCode=0 Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.261624 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd"} Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.261655 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"dd2bb0b2b1707f7496a35b8575d063bdbcbbee1e9a47279330bcbedc1e349a2c"} Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.262871 4858 status_manager.go:851] "Failed to get status for pod" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-w6t79\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.263034 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.263181 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.263569 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.650407 4858 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.166:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.651617 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.662903 4858 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/events\": dial tcp 38.102.83.166:6443: connect: connection refused" event="&Event{ObjectMeta:{auto-csr-approver-29566622-l8kbk.189e812ab4fd4072 openshift-infra 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-infra,Name:auto-csr-approver-29566622-l8kbk,UID:803f2926-2469-4a09-85ba-a1c3e4548168,APIVersion:v1,ResourceVersion:29952,FieldPath:spec.containers{oc},},Reason:Started,Message:Started container oc,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 09:02:03.655659634 +0000 UTC m=+304.976077861,LastTimestamp:2026-03-20 09:02:03.655659634 +0000 UTC m=+304.976077861,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.874397 4858 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.875099 4858 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.875848 4858 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.876186 4858 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.876560 4858 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:08 crc kubenswrapper[4858]: I0320 09:02:08.876603 4858 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 20 09:02:08 crc kubenswrapper[4858]: E0320 09:02:08.876902 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="200ms" Mar 20 09:02:09 crc kubenswrapper[4858]: E0320 09:02:09.077506 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="400ms" Mar 20 09:02:09 crc kubenswrapper[4858]: I0320 09:02:09.279751 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"4eb7db481a553862b38888ba204173378b29dc7d81f4b3d58023d37c5a769821"} Mar 20 09:02:09 crc kubenswrapper[4858]: I0320 09:02:09.279827 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cbbc58262c96350fcb3459aa065307d29f9f9a45e2ff2cb9d83c1e10bb97a24d"} Mar 20 09:02:09 crc kubenswrapper[4858]: E0320 09:02:09.280691 4858 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.166:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:09 crc kubenswrapper[4858]: I0320 09:02:09.281099 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:09 crc kubenswrapper[4858]: I0320 09:02:09.281413 4858 status_manager.go:851] "Failed to get status for pod" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-w6t79\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:09 crc kubenswrapper[4858]: I0320 09:02:09.281745 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:09 crc kubenswrapper[4858]: I0320 09:02:09.282080 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:09 crc kubenswrapper[4858]: E0320 09:02:09.480349 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="800ms" Mar 20 09:02:10 crc kubenswrapper[4858]: I0320 09:02:10.073576 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:10 crc kubenswrapper[4858]: I0320 09:02:10.073982 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:10 crc kubenswrapper[4858]: I0320 09:02:10.075379 4858 status_manager.go:851] "Failed to get status for pod" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-w6t79\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:10 crc kubenswrapper[4858]: I0320 09:02:10.075902 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:10 crc kubenswrapper[4858]: E0320 09:02:10.281214 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="1.6s" Mar 20 09:02:11 crc kubenswrapper[4858]: E0320 09:02:11.883235 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="3.2s" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.069098 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.070288 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.071030 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.071343 4858 status_manager.go:851] "Failed to get status for pod" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-w6t79\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.071714 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:15 crc kubenswrapper[4858]: E0320 09:02:15.085149 4858 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.166:6443: connect: connection refused" interval="6.4s" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.089236 4858 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.089277 4858 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:15 crc kubenswrapper[4858]: E0320 09:02:15.089771 4858 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.090476 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:15 crc kubenswrapper[4858]: W0320 09:02:15.124034 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-b9d37156453d2dfaaa433eec859a698a9aef1e3961974212b8a60454630e43b5 WatchSource:0}: Error finding container b9d37156453d2dfaaa433eec859a698a9aef1e3961974212b8a60454630e43b5: Status 404 returned error can't find the container with id b9d37156453d2dfaaa433eec859a698a9aef1e3961974212b8a60454630e43b5 Mar 20 09:02:15 crc kubenswrapper[4858]: I0320 09:02:15.327759 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b9d37156453d2dfaaa433eec859a698a9aef1e3961974212b8a60454630e43b5"} Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.339344 4858 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="9f872d34adbfa7c2275eb569848142878c24c42970903f772973a912a6722599" exitCode=0 Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.339465 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"9f872d34adbfa7c2275eb569848142878c24c42970903f772973a912a6722599"} Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.339970 4858 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.341584 4858 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.340848 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: E0320 09:02:16.342377 4858 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.342528 4858 status_manager.go:851] "Failed to get status for pod" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-w6t79\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.343153 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.343686 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.347844 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.348947 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.349034 4858 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe" exitCode=1 Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.349077 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe"} Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.349542 4858 scope.go:117] "RemoveContainer" containerID="f779320d1039dc70e2768187a349d1cd8bb3f67d803f233a4738f7b7f9112dbe" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.350173 4858 status_manager.go:851] "Failed to get status for pod" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" pod="openshift-infra/auto-csr-approver-29566622-l8kbk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/pods/auto-csr-approver-29566622-l8kbk\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.350752 4858 status_manager.go:851] "Failed to get status for pod" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" pod="openshift-marketplace/redhat-marketplace-mw84k" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-mw84k\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.351215 4858 status_manager.go:851] "Failed to get status for pod" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-w6t79\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.351790 4858 status_manager.go:851] "Failed to get status for pod" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: I0320 09:02:16.352172 4858 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.166:6443: connect: connection refused" Mar 20 09:02:16 crc kubenswrapper[4858]: E0320 09:02:16.819228 4858 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-infra/events\": dial tcp 38.102.83.166:6443: connect: connection refused" event="&Event{ObjectMeta:{auto-csr-approver-29566622-l8kbk.189e812ab4fd4072 openshift-infra 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-infra,Name:auto-csr-approver-29566622-l8kbk,UID:803f2926-2469-4a09-85ba-a1c3e4548168,APIVersion:v1,ResourceVersion:29952,FieldPath:spec.containers{oc},},Reason:Started,Message:Started container oc,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-20 09:02:03.655659634 +0000 UTC m=+304.976077861,LastTimestamp:2026-03-20 09:02:03.655659634 +0000 UTC m=+304.976077861,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 20 09:02:17 crc kubenswrapper[4858]: I0320 09:02:17.358428 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f5c85b78f123ef2972357e3b667a47abd56aba663f100f982c5aab0498d5b664"} Mar 20 09:02:17 crc kubenswrapper[4858]: I0320 09:02:17.359019 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6da190a8c01b18255b9b4d3a69fa6cea4227b9350a4ce5095446faa1ba65c894"} Mar 20 09:02:17 crc kubenswrapper[4858]: I0320 09:02:17.361231 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 20 09:02:17 crc kubenswrapper[4858]: I0320 09:02:17.362139 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 20 09:02:17 crc kubenswrapper[4858]: I0320 09:02:17.362462 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a4d325a300354e46a2990935bcef3f489de7189e0bd34474ad74851ec62c1644"} Mar 20 09:02:18 crc kubenswrapper[4858]: I0320 09:02:18.192934 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:02:18 crc kubenswrapper[4858]: I0320 09:02:18.372861 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"665bccb2bbb5a0ba2d196ec91901db5cf859047386719484d2683a1637b84d9b"} Mar 20 09:02:18 crc kubenswrapper[4858]: I0320 09:02:18.372933 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9becf3fa30bf1c3168b27a6a7768d4606d8e56ac45b2ca271ecf9d3433c91f17"} Mar 20 09:02:18 crc kubenswrapper[4858]: I0320 09:02:18.811780 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:02:18 crc kubenswrapper[4858]: I0320 09:02:18.812175 4858 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 20 09:02:18 crc kubenswrapper[4858]: I0320 09:02:18.812498 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 20 09:02:19 crc kubenswrapper[4858]: I0320 09:02:19.386325 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"9d85ca6ee19668e56509e3bcd8794a93e5539bd7f8ff8597ea3bf90e4cde42c9"} Mar 20 09:02:19 crc kubenswrapper[4858]: I0320 09:02:19.386813 4858 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:19 crc kubenswrapper[4858]: I0320 09:02:19.386851 4858 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:20 crc kubenswrapper[4858]: I0320 09:02:20.090658 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:20 crc kubenswrapper[4858]: I0320 09:02:20.090752 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:20 crc kubenswrapper[4858]: I0320 09:02:20.096931 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.128651 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" podUID="a9f367b2-d0b3-4a80-933f-68bf11e63791" containerName="oauth-openshift" containerID="cri-o://740cae4298efab318dad038c9ccff9acbbd97a0a2ed3ed5670209d0c91ac352a" gracePeriod=15 Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.405568 4858 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.424806 4858 generic.go:334] "Generic (PLEG): container finished" podID="a9f367b2-d0b3-4a80-933f-68bf11e63791" containerID="740cae4298efab318dad038c9ccff9acbbd97a0a2ed3ed5670209d0c91ac352a" exitCode=0 Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.424896 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" event={"ID":"a9f367b2-d0b3-4a80-933f-68bf11e63791","Type":"ContainerDied","Data":"740cae4298efab318dad038c9ccff9acbbd97a0a2ed3ed5670209d0c91ac352a"} Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.707570 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.877176 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-cliconfig\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878758 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-service-ca\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878805 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-idp-0-file-data\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878864 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-router-certs\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878901 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-login\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878931 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtww9\" (UniqueName: \"kubernetes.io/projected/a9f367b2-d0b3-4a80-933f-68bf11e63791-kube-api-access-vtww9\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878966 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-session\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878961 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.878993 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-dir\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879028 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-ocp-branding-template\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879050 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-trusted-ca-bundle\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879097 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-provider-selection\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879130 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-serving-cert\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879191 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-policies\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879217 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-error\") pod \"a9f367b2-d0b3-4a80-933f-68bf11e63791\" (UID: \"a9f367b2-d0b3-4a80-933f-68bf11e63791\") " Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879278 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879662 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.879685 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.880300 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.881675 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.882214 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.886929 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.889268 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.889407 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.890642 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.891388 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.891545 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9f367b2-d0b3-4a80-933f-68bf11e63791-kube-api-access-vtww9" (OuterVolumeSpecName: "kube-api-access-vtww9") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "kube-api-access-vtww9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.892749 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.896587 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.902783 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a9f367b2-d0b3-4a80-933f-68bf11e63791" (UID: "a9f367b2-d0b3-4a80-933f-68bf11e63791"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981042 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981079 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981090 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtww9\" (UniqueName: \"kubernetes.io/projected/a9f367b2-d0b3-4a80-933f-68bf11e63791-kube-api-access-vtww9\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981102 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981112 4858 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981123 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981132 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981142 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981179 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981190 4858 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a9f367b2-d0b3-4a80-933f-68bf11e63791-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981200 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:24 crc kubenswrapper[4858]: I0320 09:02:24.981213 4858 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a9f367b2-d0b3-4a80-933f-68bf11e63791-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.091057 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.095668 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.099134 4858 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f3bfe6fc-de6b-48e9-80e2-238396e640b3" Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.434829 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.435041 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2c66r" event={"ID":"a9f367b2-d0b3-4a80-933f-68bf11e63791","Type":"ContainerDied","Data":"0a12e38f2d188aa3ff73040821e951e7b7c1e50600b35e2ba3035095d967c700"} Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.435081 4858 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.435120 4858 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:25 crc kubenswrapper[4858]: I0320 09:02:25.435160 4858 scope.go:117] "RemoveContainer" containerID="740cae4298efab318dad038c9ccff9acbbd97a0a2ed3ed5670209d0c91ac352a" Mar 20 09:02:26 crc kubenswrapper[4858]: I0320 09:02:26.443207 4858 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:26 crc kubenswrapper[4858]: I0320 09:02:26.443561 4858 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="ca9fdc4c-5d34-40de-bb5d-af6140462f33" Mar 20 09:02:28 crc kubenswrapper[4858]: I0320 09:02:28.818963 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:02:28 crc kubenswrapper[4858]: I0320 09:02:28.828753 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 20 09:02:30 crc kubenswrapper[4858]: I0320 09:02:30.090189 4858 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="f3bfe6fc-de6b-48e9-80e2-238396e640b3" Mar 20 09:02:33 crc kubenswrapper[4858]: I0320 09:02:33.412403 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 20 09:02:34 crc kubenswrapper[4858]: I0320 09:02:34.048292 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 20 09:02:34 crc kubenswrapper[4858]: I0320 09:02:34.112695 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 20 09:02:34 crc kubenswrapper[4858]: I0320 09:02:34.326643 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 20 09:02:34 crc kubenswrapper[4858]: I0320 09:02:34.672464 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 20 09:02:35 crc kubenswrapper[4858]: I0320 09:02:35.774672 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 20 09:02:35 crc kubenswrapper[4858]: I0320 09:02:35.840285 4858 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 20 09:02:35 crc kubenswrapper[4858]: I0320 09:02:35.853105 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mw84k","openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-2c66r"] Mar 20 09:02:35 crc kubenswrapper[4858]: I0320 09:02:35.853195 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 20 09:02:35 crc kubenswrapper[4858]: I0320 09:02:35.874250 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=11.874230413 podStartE2EDuration="11.874230413s" podCreationTimestamp="2026-03-20 09:02:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:02:35.873339756 +0000 UTC m=+337.193758003" watchObservedRunningTime="2026-03-20 09:02:35.874230413 +0000 UTC m=+337.194648610" Mar 20 09:02:35 crc kubenswrapper[4858]: I0320 09:02:35.886839 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.002939 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.073236 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.079219 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57d2d9c4-3ee7-41f7-af06-18c775cb10c4" path="/var/lib/kubelet/pods/57d2d9c4-3ee7-41f7-af06-18c775cb10c4/volumes" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.080595 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9f367b2-d0b3-4a80-933f-68bf11e63791" path="/var/lib/kubelet/pods/a9f367b2-d0b3-4a80-933f-68bf11e63791/volumes" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.104611 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.507009 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.643920 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 20 09:02:36 crc kubenswrapper[4858]: I0320 09:02:36.861285 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.067572 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.185060 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.354715 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.461551 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.482175 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.494441 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.534838 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.742479 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.764073 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.766780 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 20 09:02:37 crc kubenswrapper[4858]: I0320 09:02:37.997021 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.016974 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.134357 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.144073 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.273447 4858 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.370807 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.442288 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.493346 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.636378 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.690647 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.705533 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.757455 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.880293 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.926121 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 20 09:02:38 crc kubenswrapper[4858]: I0320 09:02:38.990579 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.053163 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.074548 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.107561 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.171929 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.197463 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.341963 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.356200 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.401844 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.434636 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.436828 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.447541 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.475024 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.482986 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.662603 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.686634 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.754255 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.780525 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.786223 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.809797 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 20 09:02:39 crc kubenswrapper[4858]: I0320 09:02:39.942931 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.013765 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.109055 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.127384 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.140936 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.143891 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.185924 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.194913 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.260743 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.448801 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.496502 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.616875 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.713557 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.725022 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.757253 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 20 09:02:40 crc kubenswrapper[4858]: I0320 09:02:40.970464 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.058979 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.085221 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.147071 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.154483 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.175915 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.257684 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.345756 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.373185 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.375108 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.408916 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.465437 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.487247 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.488443 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.500600 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.547627 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.674951 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.737784 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.755857 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c"] Mar 20 09:02:41 crc kubenswrapper[4858]: E0320 09:02:41.756078 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f367b2-d0b3-4a80-933f-68bf11e63791" containerName="oauth-openshift" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.756090 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f367b2-d0b3-4a80-933f-68bf11e63791" containerName="oauth-openshift" Mar 20 09:02:41 crc kubenswrapper[4858]: E0320 09:02:41.756111 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" containerName="oc" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.756117 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" containerName="oc" Mar 20 09:02:41 crc kubenswrapper[4858]: E0320 09:02:41.756126 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" containerName="installer" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.756133 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" containerName="installer" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.756241 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9f367b2-d0b3-4a80-933f-68bf11e63791" containerName="oauth-openshift" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.756259 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="b42b85d2-a66d-40be-af75-611eaa9d1a3a" containerName="installer" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.756269 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" containerName="oc" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.756723 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.759780 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.760145 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.760485 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.760527 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.760606 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.760564 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.760712 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.762098 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.762246 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.762131 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.762481 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.762611 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.770584 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c"] Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.782860 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.784237 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.813202 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.815046 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.850514 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.860817 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.861239 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.868633 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920071 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920141 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6492897-58f3-43e8-af2b-0101caa1eee9-audit-dir\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920182 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920407 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920538 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920583 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920620 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-session\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920650 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920672 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-audit-policies\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920695 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920795 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920871 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-login\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920940 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tcz5\" (UniqueName: \"kubernetes.io/projected/c6492897-58f3-43e8-af2b-0101caa1eee9-kube-api-access-8tcz5\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.920982 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-error\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:41 crc kubenswrapper[4858]: I0320 09:02:41.922988 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.022805 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-session\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.022907 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.022931 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-audit-policies\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024157 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-audit-policies\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024172 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024364 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024452 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-login\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024681 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tcz5\" (UniqueName: \"kubernetes.io/projected/c6492897-58f3-43e8-af2b-0101caa1eee9-kube-api-access-8tcz5\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024776 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-error\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024839 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024957 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6492897-58f3-43e8-af2b-0101caa1eee9-audit-dir\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.024992 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.025060 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.025174 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.025226 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.025612 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c6492897-58f3-43e8-af2b-0101caa1eee9-audit-dir\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.026030 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.026365 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-service-ca\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.026582 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.031818 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-login\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.032210 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.032497 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.045336 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-session\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.046010 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.046052 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.046264 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-system-router-certs\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.046177 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c6492897-58f3-43e8-af2b-0101caa1eee9-v4-0-config-user-template-error\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.049135 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tcz5\" (UniqueName: \"kubernetes.io/projected/c6492897-58f3-43e8-af2b-0101caa1eee9-kube-api-access-8tcz5\") pod \"oauth-openshift-5c4c6fc5df-d5m4c\" (UID: \"c6492897-58f3-43e8-af2b-0101caa1eee9\") " pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.075453 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.246958 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.353849 4858 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.405339 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.494706 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.520263 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c"] Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.521907 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.551578 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" event={"ID":"c6492897-58f3-43e8-af2b-0101caa1eee9","Type":"ContainerStarted","Data":"a7d2a3df9eb064f3d63fbe69b2d7f25758c0c2c24a1e5653864c8008ac456c54"} Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.572194 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.676494 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.688120 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.729443 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.766599 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.887113 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.888015 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 20 09:02:42 crc kubenswrapper[4858]: I0320 09:02:42.975588 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.053846 4858 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.056240 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.079358 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.183673 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.203258 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.359999 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.403873 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.499230 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.526920 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.528665 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.561247 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-5c4c6fc5df-d5m4c_c6492897-58f3-43e8-af2b-0101caa1eee9/oauth-openshift/0.log" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.561341 4858 generic.go:334] "Generic (PLEG): container finished" podID="c6492897-58f3-43e8-af2b-0101caa1eee9" containerID="63656aa159ddc8149187b39da9f942420413f77c2a65d413d72e24175e23de3b" exitCode=255 Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.561388 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" event={"ID":"c6492897-58f3-43e8-af2b-0101caa1eee9","Type":"ContainerDied","Data":"63656aa159ddc8149187b39da9f942420413f77c2a65d413d72e24175e23de3b"} Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.562272 4858 scope.go:117] "RemoveContainer" containerID="63656aa159ddc8149187b39da9f942420413f77c2a65d413d72e24175e23de3b" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.575934 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.578409 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.717480 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.834929 4858 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 20 09:02:43 crc kubenswrapper[4858]: I0320 09:02:43.967721 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.119849 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.288276 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.333634 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.363750 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.365035 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.371524 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.405530 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.405925 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.412628 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.443257 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.507614 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.547138 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.548149 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.548339 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.569340 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-5c4c6fc5df-d5m4c_c6492897-58f3-43e8-af2b-0101caa1eee9/oauth-openshift/1.log" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.570160 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-5c4c6fc5df-d5m4c_c6492897-58f3-43e8-af2b-0101caa1eee9/oauth-openshift/0.log" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.570218 4858 generic.go:334] "Generic (PLEG): container finished" podID="c6492897-58f3-43e8-af2b-0101caa1eee9" containerID="ad9aeba38a293fdea19714f0fc5a62551a580f46607f7f78569fee6d15182600" exitCode=255 Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.570264 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" event={"ID":"c6492897-58f3-43e8-af2b-0101caa1eee9","Type":"ContainerDied","Data":"ad9aeba38a293fdea19714f0fc5a62551a580f46607f7f78569fee6d15182600"} Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.570319 4858 scope.go:117] "RemoveContainer" containerID="63656aa159ddc8149187b39da9f942420413f77c2a65d413d72e24175e23de3b" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.571388 4858 scope.go:117] "RemoveContainer" containerID="ad9aeba38a293fdea19714f0fc5a62551a580f46607f7f78569fee6d15182600" Mar 20 09:02:44 crc kubenswrapper[4858]: E0320 09:02:44.571934 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-5c4c6fc5df-d5m4c_openshift-authentication(c6492897-58f3-43e8-af2b-0101caa1eee9)\"" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" podUID="c6492897-58f3-43e8-af2b-0101caa1eee9" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.573813 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.596341 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.623084 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.741043 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 20 09:02:44 crc kubenswrapper[4858]: I0320 09:02:44.791251 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.090503 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.093621 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.096095 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.180373 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.227620 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.361909 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.362348 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.400627 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.410103 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.561081 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.581842 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-5c4c6fc5df-d5m4c_c6492897-58f3-43e8-af2b-0101caa1eee9/oauth-openshift/1.log" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.582537 4858 scope.go:117] "RemoveContainer" containerID="ad9aeba38a293fdea19714f0fc5a62551a580f46607f7f78569fee6d15182600" Mar 20 09:02:45 crc kubenswrapper[4858]: E0320 09:02:45.582809 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-5c4c6fc5df-d5m4c_openshift-authentication(c6492897-58f3-43e8-af2b-0101caa1eee9)\"" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" podUID="c6492897-58f3-43e8-af2b-0101caa1eee9" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.659958 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.718242 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.736868 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.821831 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.862185 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.893233 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.926990 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 20 09:02:45 crc kubenswrapper[4858]: I0320 09:02:45.937884 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.079051 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.100488 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.205844 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.272582 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.346924 4858 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.347258 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://4eb7db481a553862b38888ba204173378b29dc7d81f4b3d58023d37c5a769821" gracePeriod=5 Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.392583 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.394984 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.425189 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.531857 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.540130 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.549939 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.635802 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.660531 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.675226 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.811947 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.827289 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.856957 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 20 09:02:46 crc kubenswrapper[4858]: I0320 09:02:46.988060 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.042795 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.091666 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.098766 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.131100 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.216081 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.225905 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.312376 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.394573 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.506381 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.605207 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.719254 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.785840 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.806990 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.851509 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.915642 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.938552 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 20 09:02:47 crc kubenswrapper[4858]: I0320 09:02:47.972105 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.194584 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.263506 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.275641 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.277074 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.347998 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.361839 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.460346 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.475872 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.523788 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.579532 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.638124 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.662012 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.765003 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.837867 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.863875 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 09:02:48 crc kubenswrapper[4858]: I0320 09:02:48.908586 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.027262 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.070699 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.096403 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.286685 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.319355 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.322484 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.368682 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.472934 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.542627 4858 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.592845 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.624924 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 20 09:02:49 crc kubenswrapper[4858]: I0320 09:02:49.771393 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.024949 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.086915 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.192687 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.251616 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.305464 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.382654 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.789274 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 20 09:02:50 crc kubenswrapper[4858]: I0320 09:02:50.795711 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 20 09:02:51 crc kubenswrapper[4858]: I0320 09:02:51.513917 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 20 09:02:51 crc kubenswrapper[4858]: I0320 09:02:51.641049 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 09:02:51 crc kubenswrapper[4858]: I0320 09:02:51.641110 4858 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="4eb7db481a553862b38888ba204173378b29dc7d81f4b3d58023d37c5a769821" exitCode=137 Mar 20 09:02:51 crc kubenswrapper[4858]: I0320 09:02:51.926574 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 09:02:51 crc kubenswrapper[4858]: I0320 09:02:51.927146 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.079404 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.079453 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.079607 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.079627 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.079670 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.079981 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.080019 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.080037 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.080055 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.083340 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.083389 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.084031 4858 scope.go:117] "RemoveContainer" containerID="ad9aeba38a293fdea19714f0fc5a62551a580f46607f7f78569fee6d15182600" Mar 20 09:02:52 crc kubenswrapper[4858]: E0320 09:02:52.084287 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oauth-openshift\" with CrashLoopBackOff: \"back-off 10s restarting failed container=oauth-openshift pod=oauth-openshift-5c4c6fc5df-d5m4c_openshift-authentication(c6492897-58f3-43e8-af2b-0101caa1eee9)\"" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" podUID="c6492897-58f3-43e8-af2b-0101caa1eee9" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.092412 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.106456 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.181871 4858 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.181912 4858 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.181925 4858 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.181936 4858 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.181946 4858 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.248370 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.442530 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.612981 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.648973 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.649074 4858 scope.go:117] "RemoveContainer" containerID="4eb7db481a553862b38888ba204173378b29dc7d81f4b3d58023d37c5a769821" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.649129 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 20 09:02:52 crc kubenswrapper[4858]: I0320 09:02:52.785231 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 20 09:02:54 crc kubenswrapper[4858]: I0320 09:02:54.081555 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 20 09:03:07 crc kubenswrapper[4858]: I0320 09:03:07.070440 4858 scope.go:117] "RemoveContainer" containerID="ad9aeba38a293fdea19714f0fc5a62551a580f46607f7f78569fee6d15182600" Mar 20 09:03:07 crc kubenswrapper[4858]: I0320 09:03:07.756843 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-5c4c6fc5df-d5m4c_c6492897-58f3-43e8-af2b-0101caa1eee9/oauth-openshift/1.log" Mar 20 09:03:07 crc kubenswrapper[4858]: I0320 09:03:07.757376 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" event={"ID":"c6492897-58f3-43e8-af2b-0101caa1eee9","Type":"ContainerStarted","Data":"086469d92f20d6cb5dae4a1b71013f15be45157a5cb847088b81c044caa253dd"} Mar 20 09:03:07 crc kubenswrapper[4858]: I0320 09:03:07.758163 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:03:07 crc kubenswrapper[4858]: I0320 09:03:07.764720 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" Mar 20 09:03:07 crc kubenswrapper[4858]: I0320 09:03:07.786724 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5c4c6fc5df-d5m4c" podStartSLOduration=68.786694266 podStartE2EDuration="1m8.786694266s" podCreationTimestamp="2026-03-20 09:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:03:07.782755911 +0000 UTC m=+369.103174178" watchObservedRunningTime="2026-03-20 09:03:07.786694266 +0000 UTC m=+369.107112473" Mar 20 09:03:08 crc kubenswrapper[4858]: I0320 09:03:08.767672 4858 generic.go:334] "Generic (PLEG): container finished" podID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerID="2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196" exitCode=0 Mar 20 09:03:08 crc kubenswrapper[4858]: I0320 09:03:08.767772 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" event={"ID":"80c13002-5ff3-43ea-be87-e1b2ecf4431a","Type":"ContainerDied","Data":"2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196"} Mar 20 09:03:08 crc kubenswrapper[4858]: I0320 09:03:08.768741 4858 scope.go:117] "RemoveContainer" containerID="2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196" Mar 20 09:03:09 crc kubenswrapper[4858]: I0320 09:03:09.776885 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" event={"ID":"80c13002-5ff3-43ea-be87-e1b2ecf4431a","Type":"ContainerStarted","Data":"98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d"} Mar 20 09:03:09 crc kubenswrapper[4858]: I0320 09:03:09.777693 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:03:09 crc kubenswrapper[4858]: I0320 09:03:09.780419 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:03:32 crc kubenswrapper[4858]: I0320 09:03:32.588195 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tbfmx"] Mar 20 09:03:32 crc kubenswrapper[4858]: I0320 09:03:32.589378 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tbfmx" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="registry-server" containerID="cri-o://288a6a90769ce76c5cf08631afab7f75c5b583ba545374842448f3e542b3398d" gracePeriod=2 Mar 20 09:03:32 crc kubenswrapper[4858]: I0320 09:03:32.929815 4858 generic.go:334] "Generic (PLEG): container finished" podID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerID="288a6a90769ce76c5cf08631afab7f75c5b583ba545374842448f3e542b3398d" exitCode=0 Mar 20 09:03:32 crc kubenswrapper[4858]: I0320 09:03:32.929886 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbfmx" event={"ID":"9e6ac9fc-835d-4763-a6f5-e6923a5ee981","Type":"ContainerDied","Data":"288a6a90769ce76c5cf08631afab7f75c5b583ba545374842448f3e542b3398d"} Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.525050 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.635084 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-catalog-content\") pod \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.635296 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-utilities\") pod \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.635511 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54w7b\" (UniqueName: \"kubernetes.io/projected/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-kube-api-access-54w7b\") pod \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\" (UID: \"9e6ac9fc-835d-4763-a6f5-e6923a5ee981\") " Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.636588 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-utilities" (OuterVolumeSpecName: "utilities") pod "9e6ac9fc-835d-4763-a6f5-e6923a5ee981" (UID: "9e6ac9fc-835d-4763-a6f5-e6923a5ee981"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.644547 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-kube-api-access-54w7b" (OuterVolumeSpecName: "kube-api-access-54w7b") pod "9e6ac9fc-835d-4763-a6f5-e6923a5ee981" (UID: "9e6ac9fc-835d-4763-a6f5-e6923a5ee981"). InnerVolumeSpecName "kube-api-access-54w7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.736956 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.737006 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54w7b\" (UniqueName: \"kubernetes.io/projected/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-kube-api-access-54w7b\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.771609 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e6ac9fc-835d-4763-a6f5-e6923a5ee981" (UID: "9e6ac9fc-835d-4763-a6f5-e6923a5ee981"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.838492 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e6ac9fc-835d-4763-a6f5-e6923a5ee981-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.938652 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tbfmx" event={"ID":"9e6ac9fc-835d-4763-a6f5-e6923a5ee981","Type":"ContainerDied","Data":"407758f9bb8bd82a3c7a4cc23b362dc44715169b1fd989abfdaabee06b7346bf"} Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.938732 4858 scope.go:117] "RemoveContainer" containerID="288a6a90769ce76c5cf08631afab7f75c5b583ba545374842448f3e542b3398d" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.938678 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tbfmx" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.966492 4858 scope.go:117] "RemoveContainer" containerID="9dd033e13afafac720728adc96412bc8c9145a7837a885562bb5ccd265331652" Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.984259 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tbfmx"] Mar 20 09:03:33 crc kubenswrapper[4858]: I0320 09:03:33.989942 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tbfmx"] Mar 20 09:03:34 crc kubenswrapper[4858]: I0320 09:03:34.004522 4858 scope.go:117] "RemoveContainer" containerID="d038200a11492b4bc54f1c69bef528994645f18ce99effb5afceec12cbde6d3a" Mar 20 09:03:34 crc kubenswrapper[4858]: I0320 09:03:34.079025 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" path="/var/lib/kubelet/pods/9e6ac9fc-835d-4763-a6f5-e6923a5ee981/volumes" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.146924 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566624-bswd6"] Mar 20 09:04:00 crc kubenswrapper[4858]: E0320 09:04:00.149938 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="registry-server" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.150003 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="registry-server" Mar 20 09:04:00 crc kubenswrapper[4858]: E0320 09:04:00.150045 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.150057 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 09:04:00 crc kubenswrapper[4858]: E0320 09:04:00.150073 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="extract-content" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.150082 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="extract-content" Mar 20 09:04:00 crc kubenswrapper[4858]: E0320 09:04:00.150095 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="extract-utilities" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.150104 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="extract-utilities" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.150263 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.150294 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6ac9fc-835d-4763-a6f5-e6923a5ee981" containerName="registry-server" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.150954 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-bswd6" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.155403 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-bswd6"] Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.166099 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.166190 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.166349 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.348117 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbk4r\" (UniqueName: \"kubernetes.io/projected/5eb6b203-36ba-4b13-bd48-32b7da51f525-kube-api-access-mbk4r\") pod \"auto-csr-approver-29566624-bswd6\" (UID: \"5eb6b203-36ba-4b13-bd48-32b7da51f525\") " pod="openshift-infra/auto-csr-approver-29566624-bswd6" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.449894 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbk4r\" (UniqueName: \"kubernetes.io/projected/5eb6b203-36ba-4b13-bd48-32b7da51f525-kube-api-access-mbk4r\") pod \"auto-csr-approver-29566624-bswd6\" (UID: \"5eb6b203-36ba-4b13-bd48-32b7da51f525\") " pod="openshift-infra/auto-csr-approver-29566624-bswd6" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.472508 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbk4r\" (UniqueName: \"kubernetes.io/projected/5eb6b203-36ba-4b13-bd48-32b7da51f525-kube-api-access-mbk4r\") pod \"auto-csr-approver-29566624-bswd6\" (UID: \"5eb6b203-36ba-4b13-bd48-32b7da51f525\") " pod="openshift-infra/auto-csr-approver-29566624-bswd6" Mar 20 09:04:00 crc kubenswrapper[4858]: I0320 09:04:00.486090 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-bswd6" Mar 20 09:04:01 crc kubenswrapper[4858]: I0320 09:04:00.999993 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-bswd6"] Mar 20 09:04:01 crc kubenswrapper[4858]: I0320 09:04:01.109139 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566624-bswd6" event={"ID":"5eb6b203-36ba-4b13-bd48-32b7da51f525","Type":"ContainerStarted","Data":"ea8449ceb52b51f24a666a4dcc0b5ed3963f11cb4696bca702e82e87daf9aef2"} Mar 20 09:04:03 crc kubenswrapper[4858]: I0320 09:04:03.121271 4858 generic.go:334] "Generic (PLEG): container finished" podID="5eb6b203-36ba-4b13-bd48-32b7da51f525" containerID="705f671d45368e982f7125b9811248bbe255bf1cfd970a2055f164be17a6461d" exitCode=0 Mar 20 09:04:03 crc kubenswrapper[4858]: I0320 09:04:03.121387 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566624-bswd6" event={"ID":"5eb6b203-36ba-4b13-bd48-32b7da51f525","Type":"ContainerDied","Data":"705f671d45368e982f7125b9811248bbe255bf1cfd970a2055f164be17a6461d"} Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.470054 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-bswd6" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.602474 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbk4r\" (UniqueName: \"kubernetes.io/projected/5eb6b203-36ba-4b13-bd48-32b7da51f525-kube-api-access-mbk4r\") pod \"5eb6b203-36ba-4b13-bd48-32b7da51f525\" (UID: \"5eb6b203-36ba-4b13-bd48-32b7da51f525\") " Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.610809 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eb6b203-36ba-4b13-bd48-32b7da51f525-kube-api-access-mbk4r" (OuterVolumeSpecName: "kube-api-access-mbk4r") pod "5eb6b203-36ba-4b13-bd48-32b7da51f525" (UID: "5eb6b203-36ba-4b13-bd48-32b7da51f525"). InnerVolumeSpecName "kube-api-access-mbk4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.704939 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbk4r\" (UniqueName: \"kubernetes.io/projected/5eb6b203-36ba-4b13-bd48-32b7da51f525-kube-api-access-mbk4r\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.813366 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-57gsf"] Mar 20 09:04:04 crc kubenswrapper[4858]: E0320 09:04:04.813953 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb6b203-36ba-4b13-bd48-32b7da51f525" containerName="oc" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.814024 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb6b203-36ba-4b13-bd48-32b7da51f525" containerName="oc" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.814247 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eb6b203-36ba-4b13-bd48-32b7da51f525" containerName="oc" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.814812 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.840220 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-57gsf"] Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.908880 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adac4f87-1777-41fe-a3b4-f050bee60e2a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.909220 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-bound-sa-token\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.909431 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adac4f87-1777-41fe-a3b4-f050bee60e2a-registry-certificates\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.909567 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adac4f87-1777-41fe-a3b4-f050bee60e2a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.909696 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.909811 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-registry-tls\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.909925 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj5ck\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-kube-api-access-bj5ck\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.910401 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adac4f87-1777-41fe-a3b4-f050bee60e2a-trusted-ca\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:04 crc kubenswrapper[4858]: I0320 09:04:04.937771 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.012092 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adac4f87-1777-41fe-a3b4-f050bee60e2a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.012155 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-bound-sa-token\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.012193 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adac4f87-1777-41fe-a3b4-f050bee60e2a-registry-certificates\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.012215 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adac4f87-1777-41fe-a3b4-f050bee60e2a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.012240 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-registry-tls\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.012255 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj5ck\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-kube-api-access-bj5ck\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.012335 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adac4f87-1777-41fe-a3b4-f050bee60e2a-trusted-ca\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.014255 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adac4f87-1777-41fe-a3b4-f050bee60e2a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.014570 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adac4f87-1777-41fe-a3b4-f050bee60e2a-trusted-ca\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.015273 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adac4f87-1777-41fe-a3b4-f050bee60e2a-registry-certificates\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.018419 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adac4f87-1777-41fe-a3b4-f050bee60e2a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.019221 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-registry-tls\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.031392 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-bound-sa-token\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.033747 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj5ck\" (UniqueName: \"kubernetes.io/projected/adac4f87-1777-41fe-a3b4-f050bee60e2a-kube-api-access-bj5ck\") pod \"image-registry-66df7c8f76-57gsf\" (UID: \"adac4f87-1777-41fe-a3b4-f050bee60e2a\") " pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.132829 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.135222 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566624-bswd6" event={"ID":"5eb6b203-36ba-4b13-bd48-32b7da51f525","Type":"ContainerDied","Data":"ea8449ceb52b51f24a666a4dcc0b5ed3963f11cb4696bca702e82e87daf9aef2"} Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.135259 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea8449ceb52b51f24a666a4dcc0b5ed3963f11cb4696bca702e82e87daf9aef2" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.135304 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566624-bswd6" Mar 20 09:04:05 crc kubenswrapper[4858]: I0320 09:04:05.332212 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-57gsf"] Mar 20 09:04:05 crc kubenswrapper[4858]: W0320 09:04:05.338347 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadac4f87_1777_41fe_a3b4_f050bee60e2a.slice/crio-9615b1d4a14b5af99da7ca8d12b3118cb2bbdfdc64eef07f8ef79ee374dba5ec WatchSource:0}: Error finding container 9615b1d4a14b5af99da7ca8d12b3118cb2bbdfdc64eef07f8ef79ee374dba5ec: Status 404 returned error can't find the container with id 9615b1d4a14b5af99da7ca8d12b3118cb2bbdfdc64eef07f8ef79ee374dba5ec Mar 20 09:04:06 crc kubenswrapper[4858]: I0320 09:04:06.148180 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" event={"ID":"adac4f87-1777-41fe-a3b4-f050bee60e2a","Type":"ContainerStarted","Data":"6ef55ebde6a2098e18747a42c4cd47091c4f2f95ed7ef018092cb9ed368c79f8"} Mar 20 09:04:06 crc kubenswrapper[4858]: I0320 09:04:06.148238 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" event={"ID":"adac4f87-1777-41fe-a3b4-f050bee60e2a","Type":"ContainerStarted","Data":"9615b1d4a14b5af99da7ca8d12b3118cb2bbdfdc64eef07f8ef79ee374dba5ec"} Mar 20 09:04:06 crc kubenswrapper[4858]: I0320 09:04:06.151783 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:06 crc kubenswrapper[4858]: I0320 09:04:06.175777 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" podStartSLOduration=2.175757957 podStartE2EDuration="2.175757957s" podCreationTimestamp="2026-03-20 09:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:04:06.175649394 +0000 UTC m=+427.496067611" watchObservedRunningTime="2026-03-20 09:04:06.175757957 +0000 UTC m=+427.496176154" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.553545 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dv2r"] Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.555204 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2dv2r" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="registry-server" containerID="cri-o://b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6" gracePeriod=30 Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.570584 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6kdlf"] Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.570880 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6kdlf" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="registry-server" containerID="cri-o://8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8" gracePeriod=30 Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.589599 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4s5wf"] Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.590415 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" containerID="cri-o://98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d" gracePeriod=30 Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.595833 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbnzw"] Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.596434 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pbnzw" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="registry-server" containerID="cri-o://11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392" gracePeriod=30 Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.604459 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-swvjn"] Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.605046 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-swvjn" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="registry-server" containerID="cri-o://e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31" gracePeriod=30 Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.608969 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-st6tw"] Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.609987 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.632276 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-st6tw"] Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.715891 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bbf56bc9-5bfa-4aab-8633-a596385f59a5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.715981 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6r78\" (UniqueName: \"kubernetes.io/projected/bbf56bc9-5bfa-4aab-8633-a596385f59a5-kube-api-access-z6r78\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.716048 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbf56bc9-5bfa-4aab-8633-a596385f59a5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.817776 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbf56bc9-5bfa-4aab-8633-a596385f59a5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.817861 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bbf56bc9-5bfa-4aab-8633-a596385f59a5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.817931 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6r78\" (UniqueName: \"kubernetes.io/projected/bbf56bc9-5bfa-4aab-8633-a596385f59a5-kube-api-access-z6r78\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.820354 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bbf56bc9-5bfa-4aab-8633-a596385f59a5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.825522 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bbf56bc9-5bfa-4aab-8633-a596385f59a5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:23 crc kubenswrapper[4858]: I0320 09:04:23.844850 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6r78\" (UniqueName: \"kubernetes.io/projected/bbf56bc9-5bfa-4aab-8633-a596385f59a5-kube-api-access-z6r78\") pod \"marketplace-operator-79b997595-st6tw\" (UID: \"bbf56bc9-5bfa-4aab-8633-a596385f59a5\") " pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.001098 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.018203 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.027755 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.101197 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.113456 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.121963 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-utilities\") pod \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.122033 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-catalog-content\") pod \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.122080 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-utilities\") pod \"9daba85d-2681-4f74-8094-9db79d723cee\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.122196 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-catalog-content\") pod \"9daba85d-2681-4f74-8094-9db79d723cee\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.122230 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fll4j\" (UniqueName: \"kubernetes.io/projected/9daba85d-2681-4f74-8094-9db79d723cee-kube-api-access-fll4j\") pod \"9daba85d-2681-4f74-8094-9db79d723cee\" (UID: \"9daba85d-2681-4f74-8094-9db79d723cee\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.122301 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqfcc\" (UniqueName: \"kubernetes.io/projected/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-kube-api-access-jqfcc\") pod \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\" (UID: \"e78c3dad-ee9d-4901-8c08-2db4bd2070cd\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.124272 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-utilities" (OuterVolumeSpecName: "utilities") pod "9daba85d-2681-4f74-8094-9db79d723cee" (UID: "9daba85d-2681-4f74-8094-9db79d723cee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.126002 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.126299 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-utilities" (OuterVolumeSpecName: "utilities") pod "e78c3dad-ee9d-4901-8c08-2db4bd2070cd" (UID: "e78c3dad-ee9d-4901-8c08-2db4bd2070cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.133826 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-kube-api-access-jqfcc" (OuterVolumeSpecName: "kube-api-access-jqfcc") pod "e78c3dad-ee9d-4901-8c08-2db4bd2070cd" (UID: "e78c3dad-ee9d-4901-8c08-2db4bd2070cd"). InnerVolumeSpecName "kube-api-access-jqfcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.137921 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9daba85d-2681-4f74-8094-9db79d723cee-kube-api-access-fll4j" (OuterVolumeSpecName: "kube-api-access-fll4j") pod "9daba85d-2681-4f74-8094-9db79d723cee" (UID: "9daba85d-2681-4f74-8094-9db79d723cee"). InnerVolumeSpecName "kube-api-access-fll4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.146174 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.193233 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9daba85d-2681-4f74-8094-9db79d723cee" (UID: "9daba85d-2681-4f74-8094-9db79d723cee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.226795 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-utilities\") pod \"f98a0de8-b0a6-4c33-83b9-831c88485e50\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.226860 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jczn6\" (UniqueName: \"kubernetes.io/projected/f98a0de8-b0a6-4c33-83b9-831c88485e50-kube-api-access-jczn6\") pod \"f98a0de8-b0a6-4c33-83b9-831c88485e50\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.226884 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-utilities\") pod \"508d2d5b-0a75-4130-a396-9253b685e2cd\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.226904 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-operator-metrics\") pod \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.226946 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-catalog-content\") pod \"f98a0de8-b0a6-4c33-83b9-831c88485e50\" (UID: \"f98a0de8-b0a6-4c33-83b9-831c88485e50\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227006 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pggl\" (UniqueName: \"kubernetes.io/projected/508d2d5b-0a75-4130-a396-9253b685e2cd-kube-api-access-6pggl\") pod \"508d2d5b-0a75-4130-a396-9253b685e2cd\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227054 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-catalog-content\") pod \"508d2d5b-0a75-4130-a396-9253b685e2cd\" (UID: \"508d2d5b-0a75-4130-a396-9253b685e2cd\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227105 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-trusted-ca\") pod \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227142 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldb69\" (UniqueName: \"kubernetes.io/projected/80c13002-5ff3-43ea-be87-e1b2ecf4431a-kube-api-access-ldb69\") pod \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\" (UID: \"80c13002-5ff3-43ea-be87-e1b2ecf4431a\") " Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227547 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9daba85d-2681-4f74-8094-9db79d723cee-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227561 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fll4j\" (UniqueName: \"kubernetes.io/projected/9daba85d-2681-4f74-8094-9db79d723cee-kube-api-access-fll4j\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227573 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqfcc\" (UniqueName: \"kubernetes.io/projected/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-kube-api-access-jqfcc\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227584 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227803 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-utilities" (OuterVolumeSpecName: "utilities") pod "f98a0de8-b0a6-4c33-83b9-831c88485e50" (UID: "f98a0de8-b0a6-4c33-83b9-831c88485e50"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.227849 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-utilities" (OuterVolumeSpecName: "utilities") pod "508d2d5b-0a75-4130-a396-9253b685e2cd" (UID: "508d2d5b-0a75-4130-a396-9253b685e2cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.233047 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f98a0de8-b0a6-4c33-83b9-831c88485e50-kube-api-access-jczn6" (OuterVolumeSpecName: "kube-api-access-jczn6") pod "f98a0de8-b0a6-4c33-83b9-831c88485e50" (UID: "f98a0de8-b0a6-4c33-83b9-831c88485e50"). InnerVolumeSpecName "kube-api-access-jczn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.233045 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/508d2d5b-0a75-4130-a396-9253b685e2cd-kube-api-access-6pggl" (OuterVolumeSpecName: "kube-api-access-6pggl") pod "508d2d5b-0a75-4130-a396-9253b685e2cd" (UID: "508d2d5b-0a75-4130-a396-9253b685e2cd"). InnerVolumeSpecName "kube-api-access-6pggl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.233892 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "80c13002-5ff3-43ea-be87-e1b2ecf4431a" (UID: "80c13002-5ff3-43ea-be87-e1b2ecf4431a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.236263 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "80c13002-5ff3-43ea-be87-e1b2ecf4431a" (UID: "80c13002-5ff3-43ea-be87-e1b2ecf4431a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.240714 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80c13002-5ff3-43ea-be87-e1b2ecf4431a-kube-api-access-ldb69" (OuterVolumeSpecName: "kube-api-access-ldb69") pod "80c13002-5ff3-43ea-be87-e1b2ecf4431a" (UID: "80c13002-5ff3-43ea-be87-e1b2ecf4431a"). InnerVolumeSpecName "kube-api-access-ldb69". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.242046 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e78c3dad-ee9d-4901-8c08-2db4bd2070cd" (UID: "e78c3dad-ee9d-4901-8c08-2db4bd2070cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.256611 4858 generic.go:334] "Generic (PLEG): container finished" podID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerID="e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31" exitCode=0 Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.256675 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swvjn" event={"ID":"f98a0de8-b0a6-4c33-83b9-831c88485e50","Type":"ContainerDied","Data":"e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.256703 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-swvjn" event={"ID":"f98a0de8-b0a6-4c33-83b9-831c88485e50","Type":"ContainerDied","Data":"685c6bdfa2e9045b01c50b1b66cbeb437b492e6e213b3510b2a6570519c75dfb"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.256721 4858 scope.go:117] "RemoveContainer" containerID="e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.256828 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-swvjn" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.262856 4858 generic.go:334] "Generic (PLEG): container finished" podID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerID="11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392" exitCode=0 Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.262909 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbnzw" event={"ID":"508d2d5b-0a75-4130-a396-9253b685e2cd","Type":"ContainerDied","Data":"11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.262930 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbnzw" event={"ID":"508d2d5b-0a75-4130-a396-9253b685e2cd","Type":"ContainerDied","Data":"2bfcb2274db4b700b51d133311f6673612f36e1ed3fdf10ace4e3fa5320ab875"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.262978 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbnzw" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.267669 4858 generic.go:334] "Generic (PLEG): container finished" podID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerID="98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d" exitCode=0 Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.267716 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" event={"ID":"80c13002-5ff3-43ea-be87-e1b2ecf4431a","Type":"ContainerDied","Data":"98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.267737 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" event={"ID":"80c13002-5ff3-43ea-be87-e1b2ecf4431a","Type":"ContainerDied","Data":"6a3014c91f67d42f1c355fa9bbb1be8e37cd2fb24e0ef689026a78bcdbd26be0"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.267776 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4s5wf" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.278107 4858 generic.go:334] "Generic (PLEG): container finished" podID="9daba85d-2681-4f74-8094-9db79d723cee" containerID="8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8" exitCode=0 Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.278152 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kdlf" event={"ID":"9daba85d-2681-4f74-8094-9db79d723cee","Type":"ContainerDied","Data":"8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.278198 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kdlf" event={"ID":"9daba85d-2681-4f74-8094-9db79d723cee","Type":"ContainerDied","Data":"ed4dd9dd394d72e705f8cc80dbbe1ac5b44cdde6c7c14acc4d111267aeede9d0"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.278354 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kdlf" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.294521 4858 generic.go:334] "Generic (PLEG): container finished" podID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerID="b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6" exitCode=0 Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.295557 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dv2r" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.295098 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dv2r" event={"ID":"e78c3dad-ee9d-4901-8c08-2db4bd2070cd","Type":"ContainerDied","Data":"b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.297120 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dv2r" event={"ID":"e78c3dad-ee9d-4901-8c08-2db4bd2070cd","Type":"ContainerDied","Data":"4e0d383385daed9a156bdfec4b18c063b022dd4c235f527272b8dcf6e20da973"} Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.309704 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-st6tw"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.315949 4858 scope.go:117] "RemoveContainer" containerID="1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.327439 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4s5wf"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328871 4858 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328893 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldb69\" (UniqueName: \"kubernetes.io/projected/80c13002-5ff3-43ea-be87-e1b2ecf4431a-kube-api-access-ldb69\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328906 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328919 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jczn6\" (UniqueName: \"kubernetes.io/projected/f98a0de8-b0a6-4c33-83b9-831c88485e50-kube-api-access-jczn6\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328930 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328939 4858 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/80c13002-5ff3-43ea-be87-e1b2ecf4431a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328950 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pggl\" (UniqueName: \"kubernetes.io/projected/508d2d5b-0a75-4130-a396-9253b685e2cd-kube-api-access-6pggl\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.328960 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78c3dad-ee9d-4901-8c08-2db4bd2070cd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.332629 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "508d2d5b-0a75-4130-a396-9253b685e2cd" (UID: "508d2d5b-0a75-4130-a396-9253b685e2cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.336734 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4s5wf"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.355606 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6kdlf"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.358616 4858 scope.go:117] "RemoveContainer" containerID="e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.376899 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6kdlf"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.381173 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dv2r"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.383276 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f98a0de8-b0a6-4c33-83b9-831c88485e50" (UID: "f98a0de8-b0a6-4c33-83b9-831c88485e50"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.384797 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2dv2r"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.400680 4858 scope.go:117] "RemoveContainer" containerID="e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.401341 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31\": container with ID starting with e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31 not found: ID does not exist" containerID="e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.401407 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31"} err="failed to get container status \"e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31\": rpc error: code = NotFound desc = could not find container \"e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31\": container with ID starting with e01fa821d514544a09439d7fe0dfd345517e2d3857b4513ac37c0b6f27399d31 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.401445 4858 scope.go:117] "RemoveContainer" containerID="1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.402181 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0\": container with ID starting with 1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0 not found: ID does not exist" containerID="1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.402208 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0"} err="failed to get container status \"1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0\": rpc error: code = NotFound desc = could not find container \"1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0\": container with ID starting with 1256215838e4e60261008dcbf56376bc55d403d5b99daa3ccd33f6e6fb0df7b0 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.402228 4858 scope.go:117] "RemoveContainer" containerID="e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.402622 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102\": container with ID starting with e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102 not found: ID does not exist" containerID="e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.402680 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102"} err="failed to get container status \"e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102\": rpc error: code = NotFound desc = could not find container \"e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102\": container with ID starting with e3204e2f719ec3cd0beca4e8991936a5ce360908df32ce3ad6a39720bbc90102 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.402718 4858 scope.go:117] "RemoveContainer" containerID="11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.418858 4858 scope.go:117] "RemoveContainer" containerID="33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.430564 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f98a0de8-b0a6-4c33-83b9-831c88485e50-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.430605 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/508d2d5b-0a75-4130-a396-9253b685e2cd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.448711 4858 scope.go:117] "RemoveContainer" containerID="9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.476503 4858 scope.go:117] "RemoveContainer" containerID="11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.477263 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392\": container with ID starting with 11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392 not found: ID does not exist" containerID="11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.477360 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392"} err="failed to get container status \"11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392\": rpc error: code = NotFound desc = could not find container \"11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392\": container with ID starting with 11b88a501b6dbbc61c1f61e54f79587ebc8cf4b012aced7b9d005e981b15b392 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.477400 4858 scope.go:117] "RemoveContainer" containerID="33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.477897 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13\": container with ID starting with 33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13 not found: ID does not exist" containerID="33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.477983 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13"} err="failed to get container status \"33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13\": rpc error: code = NotFound desc = could not find container \"33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13\": container with ID starting with 33ca8445a66178dd128362c2fce50a66f3e86b6803177120b13400cfa9733a13 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.478036 4858 scope.go:117] "RemoveContainer" containerID="9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.478541 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9\": container with ID starting with 9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9 not found: ID does not exist" containerID="9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.478592 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9"} err="failed to get container status \"9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9\": rpc error: code = NotFound desc = could not find container \"9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9\": container with ID starting with 9c1608ffc0e5a0ad859d8cbe85d04e73820fe9e46c1f28a296c0f9c00eaff8f9 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.478615 4858 scope.go:117] "RemoveContainer" containerID="98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.495518 4858 scope.go:117] "RemoveContainer" containerID="2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.511349 4858 scope.go:117] "RemoveContainer" containerID="98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.512252 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d\": container with ID starting with 98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d not found: ID does not exist" containerID="98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.512455 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d"} err="failed to get container status \"98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d\": rpc error: code = NotFound desc = could not find container \"98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d\": container with ID starting with 98333b26a12c540508f54c1aa3416aef9bec4c621cf18fecf7149d5c1b65ef5d not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.512513 4858 scope.go:117] "RemoveContainer" containerID="2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.513019 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196\": container with ID starting with 2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196 not found: ID does not exist" containerID="2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.513065 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196"} err="failed to get container status \"2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196\": rpc error: code = NotFound desc = could not find container \"2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196\": container with ID starting with 2c95519b8ea536e2f9f38923950e58a4eab414dd5512a9fd74d6c4891d14a196 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.513096 4858 scope.go:117] "RemoveContainer" containerID="8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.539774 4858 scope.go:117] "RemoveContainer" containerID="97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.564225 4858 scope.go:117] "RemoveContainer" containerID="d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.581792 4858 scope.go:117] "RemoveContainer" containerID="8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.582401 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8\": container with ID starting with 8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8 not found: ID does not exist" containerID="8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.582430 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8"} err="failed to get container status \"8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8\": rpc error: code = NotFound desc = could not find container \"8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8\": container with ID starting with 8bb6fde5262a36c10bd1afd036468b85afe00c50c1a8fd94dceeda58cc7b54f8 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.582455 4858 scope.go:117] "RemoveContainer" containerID="97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.582846 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4\": container with ID starting with 97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4 not found: ID does not exist" containerID="97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.582872 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4"} err="failed to get container status \"97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4\": rpc error: code = NotFound desc = could not find container \"97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4\": container with ID starting with 97ec54aeebda6b90ada3fcb62089f963415f9889dc64c63d173eb92112e4e5a4 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.582890 4858 scope.go:117] "RemoveContainer" containerID="d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.583257 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd\": container with ID starting with d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd not found: ID does not exist" containerID="d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.583307 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd"} err="failed to get container status \"d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd\": rpc error: code = NotFound desc = could not find container \"d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd\": container with ID starting with d5a26362992a8501e185a048df589c7ba83c59b0a12a7d0b82ae45a49c94dcfd not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.583404 4858 scope.go:117] "RemoveContainer" containerID="b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.600429 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-swvjn"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.604248 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-swvjn"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.623941 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbnzw"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.625705 4858 scope.go:117] "RemoveContainer" containerID="8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.628536 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbnzw"] Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.647106 4858 scope.go:117] "RemoveContainer" containerID="0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.681710 4858 scope.go:117] "RemoveContainer" containerID="b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.682440 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6\": container with ID starting with b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6 not found: ID does not exist" containerID="b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.682513 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6"} err="failed to get container status \"b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6\": rpc error: code = NotFound desc = could not find container \"b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6\": container with ID starting with b1d6fc8ee6c2b6a611e9ebf971b6cfe471e58cb5539ff5b074562834c56851f6 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.682562 4858 scope.go:117] "RemoveContainer" containerID="8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.683115 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89\": container with ID starting with 8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89 not found: ID does not exist" containerID="8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.683233 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89"} err="failed to get container status \"8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89\": rpc error: code = NotFound desc = could not find container \"8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89\": container with ID starting with 8539b7bfd9757ef5769c6882a4aa25f329dc1e7f4bd8867562822b0f1e2fcd89 not found: ID does not exist" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.683359 4858 scope.go:117] "RemoveContainer" containerID="0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca" Mar 20 09:04:24 crc kubenswrapper[4858]: E0320 09:04:24.683692 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca\": container with ID starting with 0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca not found: ID does not exist" containerID="0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca" Mar 20 09:04:24 crc kubenswrapper[4858]: I0320 09:04:24.683711 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca"} err="failed to get container status \"0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca\": rpc error: code = NotFound desc = could not find container \"0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca\": container with ID starting with 0a603fb6f55f086b2f3294ff31f3939f9575b32123b481ce5e3368e91cd695ca not found: ID does not exist" Mar 20 09:04:25 crc kubenswrapper[4858]: I0320 09:04:25.139204 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-57gsf" Mar 20 09:04:25 crc kubenswrapper[4858]: I0320 09:04:25.198112 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8crqq"] Mar 20 09:04:25 crc kubenswrapper[4858]: I0320 09:04:25.302530 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" event={"ID":"bbf56bc9-5bfa-4aab-8633-a596385f59a5","Type":"ContainerStarted","Data":"9ea93a0b9c7fb089506ee7ffdf39a920b0ddf0f2f14761ff90bbd4f825c9375c"} Mar 20 09:04:25 crc kubenswrapper[4858]: I0320 09:04:25.303283 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" event={"ID":"bbf56bc9-5bfa-4aab-8633-a596385f59a5","Type":"ContainerStarted","Data":"8ce237dc7c37e614d227c903a849fe36712233f4e694674ed80cd77b7c5124b4"} Mar 20 09:04:25 crc kubenswrapper[4858]: I0320 09:04:25.303303 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:25 crc kubenswrapper[4858]: I0320 09:04:25.306842 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" Mar 20 09:04:25 crc kubenswrapper[4858]: I0320 09:04:25.327907 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-st6tw" podStartSLOduration=2.327880115 podStartE2EDuration="2.327880115s" podCreationTimestamp="2026-03-20 09:04:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:04:25.324566923 +0000 UTC m=+446.644985140" watchObservedRunningTime="2026-03-20 09:04:25.327880115 +0000 UTC m=+446.648298312" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.078932 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" path="/var/lib/kubelet/pods/508d2d5b-0a75-4130-a396-9253b685e2cd/volumes" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.079877 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" path="/var/lib/kubelet/pods/80c13002-5ff3-43ea-be87-e1b2ecf4431a/volumes" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.080495 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9daba85d-2681-4f74-8094-9db79d723cee" path="/var/lib/kubelet/pods/9daba85d-2681-4f74-8094-9db79d723cee/volumes" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.081863 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" path="/var/lib/kubelet/pods/e78c3dad-ee9d-4901-8c08-2db4bd2070cd/volumes" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.082673 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" path="/var/lib/kubelet/pods/f98a0de8-b0a6-4c33-83b9-831c88485e50/volumes" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207488 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d65zd"] Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207684 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207696 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207707 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207717 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207725 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207731 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207743 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207749 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207761 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207767 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207776 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207782 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="extract-content" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207789 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207795 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207805 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207810 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207817 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207823 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207829 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207836 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207845 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207850 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207858 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207865 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.207874 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207880 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="extract-utilities" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207962 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="9daba85d-2681-4f74-8094-9db79d723cee" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207973 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="508d2d5b-0a75-4130-a396-9253b685e2cd" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207979 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f98a0de8-b0a6-4c33-83b9-831c88485e50" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207988 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.207999 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78c3dad-ee9d-4901-8c08-2db4bd2070cd" containerName="registry-server" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.208008 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" Mar 20 09:04:26 crc kubenswrapper[4858]: E0320 09:04:26.208100 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.208107 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="80c13002-5ff3-43ea-be87-e1b2ecf4431a" containerName="marketplace-operator" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.208752 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.212423 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.225456 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d65zd"] Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.263961 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309824-10a9-4914-bcc2-e6ec55e6da20-utilities\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.264275 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqnm2\" (UniqueName: \"kubernetes.io/projected/1b309824-10a9-4914-bcc2-e6ec55e6da20-kube-api-access-nqnm2\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.264445 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309824-10a9-4914-bcc2-e6ec55e6da20-catalog-content\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.366091 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqnm2\" (UniqueName: \"kubernetes.io/projected/1b309824-10a9-4914-bcc2-e6ec55e6da20-kube-api-access-nqnm2\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.366266 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309824-10a9-4914-bcc2-e6ec55e6da20-catalog-content\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.366851 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309824-10a9-4914-bcc2-e6ec55e6da20-utilities\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.366902 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309824-10a9-4914-bcc2-e6ec55e6da20-catalog-content\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.367206 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309824-10a9-4914-bcc2-e6ec55e6da20-utilities\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.393295 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqnm2\" (UniqueName: \"kubernetes.io/projected/1b309824-10a9-4914-bcc2-e6ec55e6da20-kube-api-access-nqnm2\") pod \"certified-operators-d65zd\" (UID: \"1b309824-10a9-4914-bcc2-e6ec55e6da20\") " pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.410179 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hkxwl"] Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.411636 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.413875 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.422060 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkxwl"] Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.469233 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqld9\" (UniqueName: \"kubernetes.io/projected/40519ad0-414d-4c1c-86f1-45ca54a1ab73-kube-api-access-tqld9\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.469350 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40519ad0-414d-4c1c-86f1-45ca54a1ab73-utilities\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.469392 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40519ad0-414d-4c1c-86f1-45ca54a1ab73-catalog-content\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.530923 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.570464 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqld9\" (UniqueName: \"kubernetes.io/projected/40519ad0-414d-4c1c-86f1-45ca54a1ab73-kube-api-access-tqld9\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.570540 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40519ad0-414d-4c1c-86f1-45ca54a1ab73-utilities\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.570576 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40519ad0-414d-4c1c-86f1-45ca54a1ab73-catalog-content\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.571144 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40519ad0-414d-4c1c-86f1-45ca54a1ab73-catalog-content\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.571406 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40519ad0-414d-4c1c-86f1-45ca54a1ab73-utilities\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.596041 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqld9\" (UniqueName: \"kubernetes.io/projected/40519ad0-414d-4c1c-86f1-45ca54a1ab73-kube-api-access-tqld9\") pod \"redhat-marketplace-hkxwl\" (UID: \"40519ad0-414d-4c1c-86f1-45ca54a1ab73\") " pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.738614 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:26 crc kubenswrapper[4858]: I0320 09:04:26.812201 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d65zd"] Mar 20 09:04:26 crc kubenswrapper[4858]: W0320 09:04:26.824502 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b309824_10a9_4914_bcc2_e6ec55e6da20.slice/crio-ee5227be9468b78bc3e0994e02cefc42defdfbd09033a392cc3b26dac5a7ad54 WatchSource:0}: Error finding container ee5227be9468b78bc3e0994e02cefc42defdfbd09033a392cc3b26dac5a7ad54: Status 404 returned error can't find the container with id ee5227be9468b78bc3e0994e02cefc42defdfbd09033a392cc3b26dac5a7ad54 Mar 20 09:04:27 crc kubenswrapper[4858]: I0320 09:04:27.188589 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkxwl"] Mar 20 09:04:27 crc kubenswrapper[4858]: W0320 09:04:27.205762 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40519ad0_414d_4c1c_86f1_45ca54a1ab73.slice/crio-566499d53ca4934f3860fc25bc06d24c115bd31804e4a3b50999866bbde2e968 WatchSource:0}: Error finding container 566499d53ca4934f3860fc25bc06d24c115bd31804e4a3b50999866bbde2e968: Status 404 returned error can't find the container with id 566499d53ca4934f3860fc25bc06d24c115bd31804e4a3b50999866bbde2e968 Mar 20 09:04:27 crc kubenswrapper[4858]: I0320 09:04:27.317202 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkxwl" event={"ID":"40519ad0-414d-4c1c-86f1-45ca54a1ab73","Type":"ContainerStarted","Data":"566499d53ca4934f3860fc25bc06d24c115bd31804e4a3b50999866bbde2e968"} Mar 20 09:04:27 crc kubenswrapper[4858]: I0320 09:04:27.319945 4858 generic.go:334] "Generic (PLEG): container finished" podID="1b309824-10a9-4914-bcc2-e6ec55e6da20" containerID="00ac91da65512c3c9102c3908f43ccbb4a55b4e2dcb14b673b7f2bd702efa8dc" exitCode=0 Mar 20 09:04:27 crc kubenswrapper[4858]: I0320 09:04:27.321638 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d65zd" event={"ID":"1b309824-10a9-4914-bcc2-e6ec55e6da20","Type":"ContainerDied","Data":"00ac91da65512c3c9102c3908f43ccbb4a55b4e2dcb14b673b7f2bd702efa8dc"} Mar 20 09:04:27 crc kubenswrapper[4858]: I0320 09:04:27.321693 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d65zd" event={"ID":"1b309824-10a9-4914-bcc2-e6ec55e6da20","Type":"ContainerStarted","Data":"ee5227be9468b78bc3e0994e02cefc42defdfbd09033a392cc3b26dac5a7ad54"} Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.326730 4858 generic.go:334] "Generic (PLEG): container finished" podID="40519ad0-414d-4c1c-86f1-45ca54a1ab73" containerID="3eddf72faa5de0a139aadea730c88542883c32865e775d9f6ebb386e371676dc" exitCode=0 Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.326993 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkxwl" event={"ID":"40519ad0-414d-4c1c-86f1-45ca54a1ab73","Type":"ContainerDied","Data":"3eddf72faa5de0a139aadea730c88542883c32865e775d9f6ebb386e371676dc"} Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.603730 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p9dvb"] Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.604985 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.613965 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.619331 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p9dvb"] Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.700900 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85554450-9564-4cfa-9410-16dad7d9a3d2-catalog-content\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.700989 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85554450-9564-4cfa-9410-16dad7d9a3d2-utilities\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.701019 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7z68\" (UniqueName: \"kubernetes.io/projected/85554450-9564-4cfa-9410-16dad7d9a3d2-kube-api-access-f7z68\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.801664 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85554450-9564-4cfa-9410-16dad7d9a3d2-catalog-content\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.802111 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85554450-9564-4cfa-9410-16dad7d9a3d2-utilities\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.802150 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7z68\" (UniqueName: \"kubernetes.io/projected/85554450-9564-4cfa-9410-16dad7d9a3d2-kube-api-access-f7z68\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.802386 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85554450-9564-4cfa-9410-16dad7d9a3d2-catalog-content\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.802669 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85554450-9564-4cfa-9410-16dad7d9a3d2-utilities\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.803230 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tnvtl"] Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.806289 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.808630 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.818088 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnvtl"] Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.869444 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7z68\" (UniqueName: \"kubernetes.io/projected/85554450-9564-4cfa-9410-16dad7d9a3d2-kube-api-access-f7z68\") pod \"redhat-operators-p9dvb\" (UID: \"85554450-9564-4cfa-9410-16dad7d9a3d2\") " pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.903462 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24t6g\" (UniqueName: \"kubernetes.io/projected/194ca060-bd99-421b-ac6d-884d999bf54d-kube-api-access-24t6g\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.903563 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194ca060-bd99-421b-ac6d-884d999bf54d-utilities\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.903594 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194ca060-bd99-421b-ac6d-884d999bf54d-catalog-content\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:28 crc kubenswrapper[4858]: I0320 09:04:28.932348 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.004470 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194ca060-bd99-421b-ac6d-884d999bf54d-utilities\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.004538 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194ca060-bd99-421b-ac6d-884d999bf54d-catalog-content\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.004573 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24t6g\" (UniqueName: \"kubernetes.io/projected/194ca060-bd99-421b-ac6d-884d999bf54d-kube-api-access-24t6g\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.004914 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/194ca060-bd99-421b-ac6d-884d999bf54d-utilities\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.005136 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/194ca060-bd99-421b-ac6d-884d999bf54d-catalog-content\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.025678 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24t6g\" (UniqueName: \"kubernetes.io/projected/194ca060-bd99-421b-ac6d-884d999bf54d-kube-api-access-24t6g\") pod \"community-operators-tnvtl\" (UID: \"194ca060-bd99-421b-ac6d-884d999bf54d\") " pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.173901 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p9dvb"] Mar 20 09:04:29 crc kubenswrapper[4858]: W0320 09:04:29.176544 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85554450_9564_4cfa_9410_16dad7d9a3d2.slice/crio-bd1ae4f1edf245aa882518b938b43f9ee20209c3dc95ecae2beda158c3bfe5a8 WatchSource:0}: Error finding container bd1ae4f1edf245aa882518b938b43f9ee20209c3dc95ecae2beda158c3bfe5a8: Status 404 returned error can't find the container with id bd1ae4f1edf245aa882518b938b43f9ee20209c3dc95ecae2beda158c3bfe5a8 Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.207973 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.352190 4858 generic.go:334] "Generic (PLEG): container finished" podID="1b309824-10a9-4914-bcc2-e6ec55e6da20" containerID="3c0dc2d1d44269ea2dad137358a50e525ef7244063971fdbe20ba37b1fcdc0d5" exitCode=0 Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.352291 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d65zd" event={"ID":"1b309824-10a9-4914-bcc2-e6ec55e6da20","Type":"ContainerDied","Data":"3c0dc2d1d44269ea2dad137358a50e525ef7244063971fdbe20ba37b1fcdc0d5"} Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.354270 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9dvb" event={"ID":"85554450-9564-4cfa-9410-16dad7d9a3d2","Type":"ContainerStarted","Data":"56cd593c6f7f2268c51876d4d01f6a0908e03cc6a397dc5d9eeb330008700e7d"} Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.354298 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9dvb" event={"ID":"85554450-9564-4cfa-9410-16dad7d9a3d2","Type":"ContainerStarted","Data":"bd1ae4f1edf245aa882518b938b43f9ee20209c3dc95ecae2beda158c3bfe5a8"} Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.364328 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkxwl" event={"ID":"40519ad0-414d-4c1c-86f1-45ca54a1ab73","Type":"ContainerStarted","Data":"1c2b64365317daa63b0499c4ad174f495ffe484341d344d1fd96e7fcc4288e25"} Mar 20 09:04:29 crc kubenswrapper[4858]: I0320 09:04:29.451422 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tnvtl"] Mar 20 09:04:30 crc kubenswrapper[4858]: I0320 09:04:30.373903 4858 generic.go:334] "Generic (PLEG): container finished" podID="194ca060-bd99-421b-ac6d-884d999bf54d" containerID="733fb6d5bbcbd6763b6123219beeee5180cfc4bcbeac9db36f9dbbcb055e3fc5" exitCode=0 Mar 20 09:04:30 crc kubenswrapper[4858]: I0320 09:04:30.374046 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvtl" event={"ID":"194ca060-bd99-421b-ac6d-884d999bf54d","Type":"ContainerDied","Data":"733fb6d5bbcbd6763b6123219beeee5180cfc4bcbeac9db36f9dbbcb055e3fc5"} Mar 20 09:04:30 crc kubenswrapper[4858]: I0320 09:04:30.374098 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvtl" event={"ID":"194ca060-bd99-421b-ac6d-884d999bf54d","Type":"ContainerStarted","Data":"b8b5c803c76643caa5c19c9e8fc45627bb286adee202e6552a2811b9d7e12a8e"} Mar 20 09:04:30 crc kubenswrapper[4858]: I0320 09:04:30.384934 4858 generic.go:334] "Generic (PLEG): container finished" podID="85554450-9564-4cfa-9410-16dad7d9a3d2" containerID="56cd593c6f7f2268c51876d4d01f6a0908e03cc6a397dc5d9eeb330008700e7d" exitCode=0 Mar 20 09:04:30 crc kubenswrapper[4858]: I0320 09:04:30.385238 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9dvb" event={"ID":"85554450-9564-4cfa-9410-16dad7d9a3d2","Type":"ContainerDied","Data":"56cd593c6f7f2268c51876d4d01f6a0908e03cc6a397dc5d9eeb330008700e7d"} Mar 20 09:04:30 crc kubenswrapper[4858]: I0320 09:04:30.388223 4858 generic.go:334] "Generic (PLEG): container finished" podID="40519ad0-414d-4c1c-86f1-45ca54a1ab73" containerID="1c2b64365317daa63b0499c4ad174f495ffe484341d344d1fd96e7fcc4288e25" exitCode=0 Mar 20 09:04:30 crc kubenswrapper[4858]: I0320 09:04:30.388262 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkxwl" event={"ID":"40519ad0-414d-4c1c-86f1-45ca54a1ab73","Type":"ContainerDied","Data":"1c2b64365317daa63b0499c4ad174f495ffe484341d344d1fd96e7fcc4288e25"} Mar 20 09:04:31 crc kubenswrapper[4858]: I0320 09:04:31.395438 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvtl" event={"ID":"194ca060-bd99-421b-ac6d-884d999bf54d","Type":"ContainerStarted","Data":"38562a86a5f83436db14fa5a2f8a203177e2a9b359c423a78260f37ecf54f1a1"} Mar 20 09:04:31 crc kubenswrapper[4858]: I0320 09:04:31.402848 4858 generic.go:334] "Generic (PLEG): container finished" podID="85554450-9564-4cfa-9410-16dad7d9a3d2" containerID="585dc18e3ba480a67c23939538c69fdc6d5c7d6bb0b17d087366311bfcf9e0b7" exitCode=0 Mar 20 09:04:31 crc kubenswrapper[4858]: I0320 09:04:31.402957 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9dvb" event={"ID":"85554450-9564-4cfa-9410-16dad7d9a3d2","Type":"ContainerDied","Data":"585dc18e3ba480a67c23939538c69fdc6d5c7d6bb0b17d087366311bfcf9e0b7"} Mar 20 09:04:31 crc kubenswrapper[4858]: I0320 09:04:31.406002 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkxwl" event={"ID":"40519ad0-414d-4c1c-86f1-45ca54a1ab73","Type":"ContainerStarted","Data":"a0607e05b7cd0ef77b73261fbf4fa705506a3594ecfe240464d0895a387e9f4e"} Mar 20 09:04:31 crc kubenswrapper[4858]: I0320 09:04:31.410596 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d65zd" event={"ID":"1b309824-10a9-4914-bcc2-e6ec55e6da20","Type":"ContainerStarted","Data":"8572aa8384151334ae6e364896236ad1b5be8789aed519974cb69c0131df3a67"} Mar 20 09:04:31 crc kubenswrapper[4858]: I0320 09:04:31.441838 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d65zd" podStartSLOduration=2.54449146 podStartE2EDuration="5.441819589s" podCreationTimestamp="2026-03-20 09:04:26 +0000 UTC" firstStartedPulling="2026-03-20 09:04:27.32351248 +0000 UTC m=+448.643930677" lastFinishedPulling="2026-03-20 09:04:30.220840589 +0000 UTC m=+451.541258806" observedRunningTime="2026-03-20 09:04:31.438091422 +0000 UTC m=+452.758509619" watchObservedRunningTime="2026-03-20 09:04:31.441819589 +0000 UTC m=+452.762237806" Mar 20 09:04:31 crc kubenswrapper[4858]: I0320 09:04:31.457745 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hkxwl" podStartSLOduration=3.301889185 podStartE2EDuration="5.457727687s" podCreationTimestamp="2026-03-20 09:04:26 +0000 UTC" firstStartedPulling="2026-03-20 09:04:28.330259589 +0000 UTC m=+449.650677786" lastFinishedPulling="2026-03-20 09:04:30.486098091 +0000 UTC m=+451.806516288" observedRunningTime="2026-03-20 09:04:31.456561253 +0000 UTC m=+452.776979450" watchObservedRunningTime="2026-03-20 09:04:31.457727687 +0000 UTC m=+452.778145884" Mar 20 09:04:32 crc kubenswrapper[4858]: I0320 09:04:32.418814 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p9dvb" event={"ID":"85554450-9564-4cfa-9410-16dad7d9a3d2","Type":"ContainerStarted","Data":"20426bf0c137d658089ded6b5c8ea61a0ce04661b94f278ed483b8aa8147d968"} Mar 20 09:04:32 crc kubenswrapper[4858]: I0320 09:04:32.421466 4858 generic.go:334] "Generic (PLEG): container finished" podID="194ca060-bd99-421b-ac6d-884d999bf54d" containerID="38562a86a5f83436db14fa5a2f8a203177e2a9b359c423a78260f37ecf54f1a1" exitCode=0 Mar 20 09:04:32 crc kubenswrapper[4858]: I0320 09:04:32.421549 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvtl" event={"ID":"194ca060-bd99-421b-ac6d-884d999bf54d","Type":"ContainerDied","Data":"38562a86a5f83436db14fa5a2f8a203177e2a9b359c423a78260f37ecf54f1a1"} Mar 20 09:04:32 crc kubenswrapper[4858]: I0320 09:04:32.445339 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p9dvb" podStartSLOduration=1.899276733 podStartE2EDuration="4.445303511s" podCreationTimestamp="2026-03-20 09:04:28 +0000 UTC" firstStartedPulling="2026-03-20 09:04:29.359560575 +0000 UTC m=+450.679978772" lastFinishedPulling="2026-03-20 09:04:31.905587363 +0000 UTC m=+453.226005550" observedRunningTime="2026-03-20 09:04:32.441743129 +0000 UTC m=+453.762161326" watchObservedRunningTime="2026-03-20 09:04:32.445303511 +0000 UTC m=+453.765721708" Mar 20 09:04:33 crc kubenswrapper[4858]: I0320 09:04:33.429418 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tnvtl" event={"ID":"194ca060-bd99-421b-ac6d-884d999bf54d","Type":"ContainerStarted","Data":"d791e61bec057b23e1b8d64e414c5a001cf30fe53b5a624dcb4b136f013db14f"} Mar 20 09:04:33 crc kubenswrapper[4858]: I0320 09:04:33.451896 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tnvtl" podStartSLOduration=2.737096624 podStartE2EDuration="5.451878133s" podCreationTimestamp="2026-03-20 09:04:28 +0000 UTC" firstStartedPulling="2026-03-20 09:04:30.380330858 +0000 UTC m=+451.700749075" lastFinishedPulling="2026-03-20 09:04:33.095112387 +0000 UTC m=+454.415530584" observedRunningTime="2026-03-20 09:04:33.448856046 +0000 UTC m=+454.769274273" watchObservedRunningTime="2026-03-20 09:04:33.451878133 +0000 UTC m=+454.772296330" Mar 20 09:04:36 crc kubenswrapper[4858]: I0320 09:04:36.531399 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:36 crc kubenswrapper[4858]: I0320 09:04:36.531869 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:36 crc kubenswrapper[4858]: I0320 09:04:36.579649 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:36 crc kubenswrapper[4858]: I0320 09:04:36.738897 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:36 crc kubenswrapper[4858]: I0320 09:04:36.739186 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:36 crc kubenswrapper[4858]: I0320 09:04:36.782155 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:37 crc kubenswrapper[4858]: I0320 09:04:37.499959 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hkxwl" Mar 20 09:04:37 crc kubenswrapper[4858]: I0320 09:04:37.890152 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:04:37 crc kubenswrapper[4858]: I0320 09:04:37.890226 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:04:37 crc kubenswrapper[4858]: I0320 09:04:37.890563 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d65zd" Mar 20 09:04:38 crc kubenswrapper[4858]: I0320 09:04:38.933935 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:38 crc kubenswrapper[4858]: I0320 09:04:38.934033 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:38 crc kubenswrapper[4858]: I0320 09:04:38.977127 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:39 crc kubenswrapper[4858]: I0320 09:04:39.208896 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:39 crc kubenswrapper[4858]: I0320 09:04:39.208960 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:39 crc kubenswrapper[4858]: I0320 09:04:39.266582 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:39 crc kubenswrapper[4858]: I0320 09:04:39.505814 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p9dvb" Mar 20 09:04:39 crc kubenswrapper[4858]: I0320 09:04:39.518240 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tnvtl" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.236724 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" podUID="c2e4d497-a390-4102-961e-8334641b8867" containerName="registry" containerID="cri-o://11963e8bc58b0b524149f0be133e60054814a9211fbbd86007e04d09ebdc89ca" gracePeriod=30 Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.553927 4858 generic.go:334] "Generic (PLEG): container finished" podID="c2e4d497-a390-4102-961e-8334641b8867" containerID="11963e8bc58b0b524149f0be133e60054814a9211fbbd86007e04d09ebdc89ca" exitCode=0 Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.554029 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" event={"ID":"c2e4d497-a390-4102-961e-8334641b8867","Type":"ContainerDied","Data":"11963e8bc58b0b524149f0be133e60054814a9211fbbd86007e04d09ebdc89ca"} Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.669066 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.863566 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-bound-sa-token\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.863641 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-registry-tls\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.863668 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-registry-certificates\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.863689 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-trusted-ca\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.863732 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x79ww\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-kube-api-access-x79ww\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.863825 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2e4d497-a390-4102-961e-8334641b8867-ca-trust-extracted\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.863902 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2e4d497-a390-4102-961e-8334641b8867-installation-pull-secrets\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.864048 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"c2e4d497-a390-4102-961e-8334641b8867\" (UID: \"c2e4d497-a390-4102-961e-8334641b8867\") " Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.880159 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.880221 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-kube-api-access-x79ww" (OuterVolumeSpecName: "kube-api-access-x79ww") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "kube-api-access-x79ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.881970 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.896958 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2e4d497-a390-4102-961e-8334641b8867-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.899138 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.911719 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.923264 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2e4d497-a390-4102-961e-8334641b8867-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.924944 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "c2e4d497-a390-4102-961e-8334641b8867" (UID: "c2e4d497-a390-4102-961e-8334641b8867"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.965363 4858 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/c2e4d497-a390-4102-961e-8334641b8867-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.965428 4858 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.965441 4858 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.965451 4858 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.965461 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c2e4d497-a390-4102-961e-8334641b8867-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.965472 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x79ww\" (UniqueName: \"kubernetes.io/projected/c2e4d497-a390-4102-961e-8334641b8867-kube-api-access-x79ww\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:50 crc kubenswrapper[4858]: I0320 09:04:50.965483 4858 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/c2e4d497-a390-4102-961e-8334641b8867-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 20 09:04:51 crc kubenswrapper[4858]: I0320 09:04:51.561799 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" event={"ID":"c2e4d497-a390-4102-961e-8334641b8867","Type":"ContainerDied","Data":"262a64e95dff06f5ab981c81882b0c82fd7e84f143bc1093af68dffdde952bf9"} Mar 20 09:04:51 crc kubenswrapper[4858]: I0320 09:04:51.561872 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8crqq" Mar 20 09:04:51 crc kubenswrapper[4858]: I0320 09:04:51.561883 4858 scope.go:117] "RemoveContainer" containerID="11963e8bc58b0b524149f0be133e60054814a9211fbbd86007e04d09ebdc89ca" Mar 20 09:04:51 crc kubenswrapper[4858]: I0320 09:04:51.595526 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8crqq"] Mar 20 09:04:51 crc kubenswrapper[4858]: I0320 09:04:51.600103 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8crqq"] Mar 20 09:04:52 crc kubenswrapper[4858]: I0320 09:04:52.079004 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2e4d497-a390-4102-961e-8334641b8867" path="/var/lib/kubelet/pods/c2e4d497-a390-4102-961e-8334641b8867/volumes" Mar 20 09:05:07 crc kubenswrapper[4858]: I0320 09:05:07.890914 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:05:07 crc kubenswrapper[4858]: I0320 09:05:07.891650 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:05:37 crc kubenswrapper[4858]: I0320 09:05:37.889838 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:05:37 crc kubenswrapper[4858]: I0320 09:05:37.890630 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:05:37 crc kubenswrapper[4858]: I0320 09:05:37.890710 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:05:37 crc kubenswrapper[4858]: I0320 09:05:37.891511 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd2bb0b2b1707f7496a35b8575d063bdbcbbee1e9a47279330bcbedc1e349a2c"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:05:37 crc kubenswrapper[4858]: I0320 09:05:37.891581 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://dd2bb0b2b1707f7496a35b8575d063bdbcbbee1e9a47279330bcbedc1e349a2c" gracePeriod=600 Mar 20 09:05:38 crc kubenswrapper[4858]: I0320 09:05:38.924956 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="dd2bb0b2b1707f7496a35b8575d063bdbcbbee1e9a47279330bcbedc1e349a2c" exitCode=0 Mar 20 09:05:38 crc kubenswrapper[4858]: I0320 09:05:38.925012 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"dd2bb0b2b1707f7496a35b8575d063bdbcbbee1e9a47279330bcbedc1e349a2c"} Mar 20 09:05:38 crc kubenswrapper[4858]: I0320 09:05:38.925736 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"3e6751ee8d22e07ea1e61646d6ebfd0426280178cdcbe58bc656568144848749"} Mar 20 09:05:38 crc kubenswrapper[4858]: I0320 09:05:38.925759 4858 scope.go:117] "RemoveContainer" containerID="d9fa2c2b0c5d1cc21f952949dff40c23a8453d520a888fc83d6248336f0e15cd" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.155379 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566626-s979k"] Mar 20 09:06:00 crc kubenswrapper[4858]: E0320 09:06:00.156783 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2e4d497-a390-4102-961e-8334641b8867" containerName="registry" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.156799 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2e4d497-a390-4102-961e-8334641b8867" containerName="registry" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.156914 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2e4d497-a390-4102-961e-8334641b8867" containerName="registry" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.157370 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-s979k" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.164176 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.164869 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.165091 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.169452 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-s979k"] Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.215577 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlmmm\" (UniqueName: \"kubernetes.io/projected/a8b8a11e-6e3c-4d93-9d52-46fe767e7b07-kube-api-access-zlmmm\") pod \"auto-csr-approver-29566626-s979k\" (UID: \"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07\") " pod="openshift-infra/auto-csr-approver-29566626-s979k" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.316245 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlmmm\" (UniqueName: \"kubernetes.io/projected/a8b8a11e-6e3c-4d93-9d52-46fe767e7b07-kube-api-access-zlmmm\") pod \"auto-csr-approver-29566626-s979k\" (UID: \"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07\") " pod="openshift-infra/auto-csr-approver-29566626-s979k" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.351129 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlmmm\" (UniqueName: \"kubernetes.io/projected/a8b8a11e-6e3c-4d93-9d52-46fe767e7b07-kube-api-access-zlmmm\") pod \"auto-csr-approver-29566626-s979k\" (UID: \"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07\") " pod="openshift-infra/auto-csr-approver-29566626-s979k" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.487239 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-s979k" Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.723527 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-s979k"] Mar 20 09:06:00 crc kubenswrapper[4858]: W0320 09:06:00.733612 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8b8a11e_6e3c_4d93_9d52_46fe767e7b07.slice/crio-9efdb8d7b50e8d5f502080f40e8f1d716eb40c66bbce46c3fecc4294bb611e2f WatchSource:0}: Error finding container 9efdb8d7b50e8d5f502080f40e8f1d716eb40c66bbce46c3fecc4294bb611e2f: Status 404 returned error can't find the container with id 9efdb8d7b50e8d5f502080f40e8f1d716eb40c66bbce46c3fecc4294bb611e2f Mar 20 09:06:00 crc kubenswrapper[4858]: I0320 09:06:00.739333 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:06:01 crc kubenswrapper[4858]: I0320 09:06:01.102656 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566626-s979k" event={"ID":"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07","Type":"ContainerStarted","Data":"9efdb8d7b50e8d5f502080f40e8f1d716eb40c66bbce46c3fecc4294bb611e2f"} Mar 20 09:06:02 crc kubenswrapper[4858]: I0320 09:06:02.109772 4858 generic.go:334] "Generic (PLEG): container finished" podID="a8b8a11e-6e3c-4d93-9d52-46fe767e7b07" containerID="1c4b5289f1779989ca82b5eeca4aeaebf2a62ae62c406295be9f3b6e5dabeb23" exitCode=0 Mar 20 09:06:02 crc kubenswrapper[4858]: I0320 09:06:02.109933 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566626-s979k" event={"ID":"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07","Type":"ContainerDied","Data":"1c4b5289f1779989ca82b5eeca4aeaebf2a62ae62c406295be9f3b6e5dabeb23"} Mar 20 09:06:03 crc kubenswrapper[4858]: I0320 09:06:03.334755 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-s979k" Mar 20 09:06:03 crc kubenswrapper[4858]: I0320 09:06:03.359566 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlmmm\" (UniqueName: \"kubernetes.io/projected/a8b8a11e-6e3c-4d93-9d52-46fe767e7b07-kube-api-access-zlmmm\") pod \"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07\" (UID: \"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07\") " Mar 20 09:06:03 crc kubenswrapper[4858]: I0320 09:06:03.366649 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8b8a11e-6e3c-4d93-9d52-46fe767e7b07-kube-api-access-zlmmm" (OuterVolumeSpecName: "kube-api-access-zlmmm") pod "a8b8a11e-6e3c-4d93-9d52-46fe767e7b07" (UID: "a8b8a11e-6e3c-4d93-9d52-46fe767e7b07"). InnerVolumeSpecName "kube-api-access-zlmmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:06:03 crc kubenswrapper[4858]: I0320 09:06:03.461670 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlmmm\" (UniqueName: \"kubernetes.io/projected/a8b8a11e-6e3c-4d93-9d52-46fe767e7b07-kube-api-access-zlmmm\") on node \"crc\" DevicePath \"\"" Mar 20 09:06:04 crc kubenswrapper[4858]: I0320 09:06:04.125747 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566626-s979k" event={"ID":"a8b8a11e-6e3c-4d93-9d52-46fe767e7b07","Type":"ContainerDied","Data":"9efdb8d7b50e8d5f502080f40e8f1d716eb40c66bbce46c3fecc4294bb611e2f"} Mar 20 09:06:04 crc kubenswrapper[4858]: I0320 09:06:04.125803 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9efdb8d7b50e8d5f502080f40e8f1d716eb40c66bbce46c3fecc4294bb611e2f" Mar 20 09:06:04 crc kubenswrapper[4858]: I0320 09:06:04.125836 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566626-s979k" Mar 20 09:06:04 crc kubenswrapper[4858]: I0320 09:06:04.405518 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-gz4tx"] Mar 20 09:06:04 crc kubenswrapper[4858]: I0320 09:06:04.412810 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566620-gz4tx"] Mar 20 09:06:06 crc kubenswrapper[4858]: I0320 09:06:06.078191 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74fe10ec-a162-4c93-b2d3-1a80745e7fcc" path="/var/lib/kubelet/pods/74fe10ec-a162-4c93-b2d3-1a80745e7fcc/volumes" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.184971 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566628-54mc2"] Mar 20 09:08:00 crc kubenswrapper[4858]: E0320 09:08:00.186234 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8b8a11e-6e3c-4d93-9d52-46fe767e7b07" containerName="oc" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.186251 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8b8a11e-6e3c-4d93-9d52-46fe767e7b07" containerName="oc" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.186399 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8b8a11e-6e3c-4d93-9d52-46fe767e7b07" containerName="oc" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.186946 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-54mc2" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.189233 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.189935 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-54mc2"] Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.190454 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.190514 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:08:00 crc kubenswrapper[4858]: I0320 09:08:00.231963 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6whcv\" (UniqueName: \"kubernetes.io/projected/7f99eb15-1e3e-4ed9-932d-991b29255c03-kube-api-access-6whcv\") pod \"auto-csr-approver-29566628-54mc2\" (UID: \"7f99eb15-1e3e-4ed9-932d-991b29255c03\") " pod="openshift-infra/auto-csr-approver-29566628-54mc2" Mar 20 09:08:01 crc kubenswrapper[4858]: I0320 09:08:00.333262 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6whcv\" (UniqueName: \"kubernetes.io/projected/7f99eb15-1e3e-4ed9-932d-991b29255c03-kube-api-access-6whcv\") pod \"auto-csr-approver-29566628-54mc2\" (UID: \"7f99eb15-1e3e-4ed9-932d-991b29255c03\") " pod="openshift-infra/auto-csr-approver-29566628-54mc2" Mar 20 09:08:01 crc kubenswrapper[4858]: I0320 09:08:01.256629 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6whcv\" (UniqueName: \"kubernetes.io/projected/7f99eb15-1e3e-4ed9-932d-991b29255c03-kube-api-access-6whcv\") pod \"auto-csr-approver-29566628-54mc2\" (UID: \"7f99eb15-1e3e-4ed9-932d-991b29255c03\") " pod="openshift-infra/auto-csr-approver-29566628-54mc2" Mar 20 09:08:01 crc kubenswrapper[4858]: I0320 09:08:01.409929 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-54mc2" Mar 20 09:08:01 crc kubenswrapper[4858]: I0320 09:08:01.629459 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-54mc2"] Mar 20 09:08:02 crc kubenswrapper[4858]: I0320 09:08:02.253246 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566628-54mc2" event={"ID":"7f99eb15-1e3e-4ed9-932d-991b29255c03","Type":"ContainerStarted","Data":"6d7709446251f7be887bb7819e6a053af1a4a80319f27b3bdb108d0753f69756"} Mar 20 09:08:04 crc kubenswrapper[4858]: I0320 09:08:04.272276 4858 generic.go:334] "Generic (PLEG): container finished" podID="7f99eb15-1e3e-4ed9-932d-991b29255c03" containerID="e8534a5e53e6fda3045f827a8b7a484525243445c73e3763e967a16eaa46ebf0" exitCode=0 Mar 20 09:08:04 crc kubenswrapper[4858]: I0320 09:08:04.272382 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566628-54mc2" event={"ID":"7f99eb15-1e3e-4ed9-932d-991b29255c03","Type":"ContainerDied","Data":"e8534a5e53e6fda3045f827a8b7a484525243445c73e3763e967a16eaa46ebf0"} Mar 20 09:08:05 crc kubenswrapper[4858]: I0320 09:08:05.599775 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-54mc2" Mar 20 09:08:05 crc kubenswrapper[4858]: I0320 09:08:05.710657 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6whcv\" (UniqueName: \"kubernetes.io/projected/7f99eb15-1e3e-4ed9-932d-991b29255c03-kube-api-access-6whcv\") pod \"7f99eb15-1e3e-4ed9-932d-991b29255c03\" (UID: \"7f99eb15-1e3e-4ed9-932d-991b29255c03\") " Mar 20 09:08:05 crc kubenswrapper[4858]: I0320 09:08:05.719075 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f99eb15-1e3e-4ed9-932d-991b29255c03-kube-api-access-6whcv" (OuterVolumeSpecName: "kube-api-access-6whcv") pod "7f99eb15-1e3e-4ed9-932d-991b29255c03" (UID: "7f99eb15-1e3e-4ed9-932d-991b29255c03"). InnerVolumeSpecName "kube-api-access-6whcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:08:05 crc kubenswrapper[4858]: I0320 09:08:05.813278 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6whcv\" (UniqueName: \"kubernetes.io/projected/7f99eb15-1e3e-4ed9-932d-991b29255c03-kube-api-access-6whcv\") on node \"crc\" DevicePath \"\"" Mar 20 09:08:06 crc kubenswrapper[4858]: I0320 09:08:06.287687 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566628-54mc2" event={"ID":"7f99eb15-1e3e-4ed9-932d-991b29255c03","Type":"ContainerDied","Data":"6d7709446251f7be887bb7819e6a053af1a4a80319f27b3bdb108d0753f69756"} Mar 20 09:08:06 crc kubenswrapper[4858]: I0320 09:08:06.287738 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d7709446251f7be887bb7819e6a053af1a4a80319f27b3bdb108d0753f69756" Mar 20 09:08:06 crc kubenswrapper[4858]: I0320 09:08:06.287796 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566628-54mc2" Mar 20 09:08:06 crc kubenswrapper[4858]: I0320 09:08:06.672783 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-l8kbk"] Mar 20 09:08:06 crc kubenswrapper[4858]: I0320 09:08:06.675845 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566622-l8kbk"] Mar 20 09:08:07 crc kubenswrapper[4858]: I0320 09:08:07.890610 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:08:07 crc kubenswrapper[4858]: I0320 09:08:07.891210 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:08:08 crc kubenswrapper[4858]: I0320 09:08:08.078641 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="803f2926-2469-4a09-85ba-a1c3e4548168" path="/var/lib/kubelet/pods/803f2926-2469-4a09-85ba-a1c3e4548168/volumes" Mar 20 09:08:25 crc kubenswrapper[4858]: I0320 09:08:25.179941 4858 scope.go:117] "RemoveContainer" containerID="697e39d85e38f417fa41aff2272b77f57de149bb17df404e92c2553d9fc17a58" Mar 20 09:08:25 crc kubenswrapper[4858]: I0320 09:08:25.223032 4858 scope.go:117] "RemoveContainer" containerID="28ba9dd114777b6309024ecd2a663cdd2b72f0255cf4c838204de1a5503439e8" Mar 20 09:08:37 crc kubenswrapper[4858]: I0320 09:08:37.889799 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:08:37 crc kubenswrapper[4858]: I0320 09:08:37.890525 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:09:07 crc kubenswrapper[4858]: I0320 09:09:07.891073 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:09:07 crc kubenswrapper[4858]: I0320 09:09:07.892199 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:09:07 crc kubenswrapper[4858]: I0320 09:09:07.892286 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:09:07 crc kubenswrapper[4858]: I0320 09:09:07.893557 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e6751ee8d22e07ea1e61646d6ebfd0426280178cdcbe58bc656568144848749"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:09:07 crc kubenswrapper[4858]: I0320 09:09:07.893665 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://3e6751ee8d22e07ea1e61646d6ebfd0426280178cdcbe58bc656568144848749" gracePeriod=600 Mar 20 09:09:08 crc kubenswrapper[4858]: I0320 09:09:08.709532 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="3e6751ee8d22e07ea1e61646d6ebfd0426280178cdcbe58bc656568144848749" exitCode=0 Mar 20 09:09:08 crc kubenswrapper[4858]: I0320 09:09:08.709727 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"3e6751ee8d22e07ea1e61646d6ebfd0426280178cdcbe58bc656568144848749"} Mar 20 09:09:08 crc kubenswrapper[4858]: I0320 09:09:08.710427 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"2e450803b001f2a0183f7a90ae4b9b24f8c995b72aa498eab30eafb0ce280f7d"} Mar 20 09:09:08 crc kubenswrapper[4858]: I0320 09:09:08.710451 4858 scope.go:117] "RemoveContainer" containerID="dd2bb0b2b1707f7496a35b8575d063bdbcbbee1e9a47279330bcbedc1e349a2c" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.136926 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566630-lp6j8"] Mar 20 09:10:00 crc kubenswrapper[4858]: E0320 09:10:00.137997 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f99eb15-1e3e-4ed9-932d-991b29255c03" containerName="oc" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.138012 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f99eb15-1e3e-4ed9-932d-991b29255c03" containerName="oc" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.138122 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f99eb15-1e3e-4ed9-932d-991b29255c03" containerName="oc" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.138603 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-lp6j8" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.140872 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.141046 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.141153 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.146178 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-lp6j8"] Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.169838 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6vrz\" (UniqueName: \"kubernetes.io/projected/63e4e001-1a9c-4669-b56a-d0d8abfe327d-kube-api-access-d6vrz\") pod \"auto-csr-approver-29566630-lp6j8\" (UID: \"63e4e001-1a9c-4669-b56a-d0d8abfe327d\") " pod="openshift-infra/auto-csr-approver-29566630-lp6j8" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.271818 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6vrz\" (UniqueName: \"kubernetes.io/projected/63e4e001-1a9c-4669-b56a-d0d8abfe327d-kube-api-access-d6vrz\") pod \"auto-csr-approver-29566630-lp6j8\" (UID: \"63e4e001-1a9c-4669-b56a-d0d8abfe327d\") " pod="openshift-infra/auto-csr-approver-29566630-lp6j8" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.291844 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6vrz\" (UniqueName: \"kubernetes.io/projected/63e4e001-1a9c-4669-b56a-d0d8abfe327d-kube-api-access-d6vrz\") pod \"auto-csr-approver-29566630-lp6j8\" (UID: \"63e4e001-1a9c-4669-b56a-d0d8abfe327d\") " pod="openshift-infra/auto-csr-approver-29566630-lp6j8" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.456763 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-lp6j8" Mar 20 09:10:00 crc kubenswrapper[4858]: I0320 09:10:00.709226 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-lp6j8"] Mar 20 09:10:01 crc kubenswrapper[4858]: I0320 09:10:01.063899 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566630-lp6j8" event={"ID":"63e4e001-1a9c-4669-b56a-d0d8abfe327d","Type":"ContainerStarted","Data":"2db2bcfc104c32bc6f465d969d0db2c7ef15f5e859373cd39fc9d0f75cf4c2dc"} Mar 20 09:10:03 crc kubenswrapper[4858]: I0320 09:10:03.080172 4858 generic.go:334] "Generic (PLEG): container finished" podID="63e4e001-1a9c-4669-b56a-d0d8abfe327d" containerID="006663afa8696a0b72fbe29cee51b6e695e4265a339622aaebf05e1b3da841e6" exitCode=0 Mar 20 09:10:03 crc kubenswrapper[4858]: I0320 09:10:03.080289 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566630-lp6j8" event={"ID":"63e4e001-1a9c-4669-b56a-d0d8abfe327d","Type":"ContainerDied","Data":"006663afa8696a0b72fbe29cee51b6e695e4265a339622aaebf05e1b3da841e6"} Mar 20 09:10:04 crc kubenswrapper[4858]: I0320 09:10:04.292846 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-lp6j8" Mar 20 09:10:04 crc kubenswrapper[4858]: I0320 09:10:04.330716 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6vrz\" (UniqueName: \"kubernetes.io/projected/63e4e001-1a9c-4669-b56a-d0d8abfe327d-kube-api-access-d6vrz\") pod \"63e4e001-1a9c-4669-b56a-d0d8abfe327d\" (UID: \"63e4e001-1a9c-4669-b56a-d0d8abfe327d\") " Mar 20 09:10:04 crc kubenswrapper[4858]: I0320 09:10:04.337897 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e4e001-1a9c-4669-b56a-d0d8abfe327d-kube-api-access-d6vrz" (OuterVolumeSpecName: "kube-api-access-d6vrz") pod "63e4e001-1a9c-4669-b56a-d0d8abfe327d" (UID: "63e4e001-1a9c-4669-b56a-d0d8abfe327d"). InnerVolumeSpecName "kube-api-access-d6vrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:10:04 crc kubenswrapper[4858]: I0320 09:10:04.433302 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6vrz\" (UniqueName: \"kubernetes.io/projected/63e4e001-1a9c-4669-b56a-d0d8abfe327d-kube-api-access-d6vrz\") on node \"crc\" DevicePath \"\"" Mar 20 09:10:05 crc kubenswrapper[4858]: I0320 09:10:05.098112 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566630-lp6j8" event={"ID":"63e4e001-1a9c-4669-b56a-d0d8abfe327d","Type":"ContainerDied","Data":"2db2bcfc104c32bc6f465d969d0db2c7ef15f5e859373cd39fc9d0f75cf4c2dc"} Mar 20 09:10:05 crc kubenswrapper[4858]: I0320 09:10:05.098481 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2db2bcfc104c32bc6f465d969d0db2c7ef15f5e859373cd39fc9d0f75cf4c2dc" Mar 20 09:10:05 crc kubenswrapper[4858]: I0320 09:10:05.098204 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566630-lp6j8" Mar 20 09:10:05 crc kubenswrapper[4858]: I0320 09:10:05.355056 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-bswd6"] Mar 20 09:10:05 crc kubenswrapper[4858]: I0320 09:10:05.358105 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566624-bswd6"] Mar 20 09:10:06 crc kubenswrapper[4858]: I0320 09:10:06.082900 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eb6b203-36ba-4b13-bd48-32b7da51f525" path="/var/lib/kubelet/pods/5eb6b203-36ba-4b13-bd48-32b7da51f525/volumes" Mar 20 09:10:25 crc kubenswrapper[4858]: I0320 09:10:25.301657 4858 scope.go:117] "RemoveContainer" containerID="705f671d45368e982f7125b9811248bbe255bf1cfd970a2055f164be17a6461d" Mar 20 09:10:46 crc kubenswrapper[4858]: I0320 09:10:46.289830 4858 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 20 09:11:37 crc kubenswrapper[4858]: I0320 09:11:37.890377 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:11:37 crc kubenswrapper[4858]: I0320 09:11:37.891386 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.379750 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dwpzf"] Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.381040 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-controller" containerID="cri-o://dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.381098 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="nbdb" containerID="cri-o://94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.381165 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.381252 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="northd" containerID="cri-o://892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.381272 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-node" containerID="cri-o://69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.381283 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-acl-logging" containerID="cri-o://5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.381178 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="sbdb" containerID="cri-o://d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.418237 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" containerID="cri-o://96b99b2166595d06311919f39bcf5f3bcd3dd03439156ccbcfd3b92bcdf473f0" gracePeriod=30 Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.572733 4858 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24656c62_314b_4c20_adf1_217d58a95f57.slice/crio-ba981e9a0ce9b14170b3eabfd0ccf4d14c784ab266253c7e4f5a08575832c5ac.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21fd7c33_ddc7_4a05_a922_472eb8ccd4e1.slice/crio-conmon-69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21fd7c33_ddc7_4a05_a922_472eb8ccd4e1.slice/crio-conmon-cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24656c62_314b_4c20_adf1_217d58a95f57.slice/crio-conmon-ba981e9a0ce9b14170b3eabfd0ccf4d14c784ab266253c7e4f5a08575832c5ac.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21fd7c33_ddc7_4a05_a922_472eb8ccd4e1.slice/crio-conmon-dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21fd7c33_ddc7_4a05_a922_472eb8ccd4e1.slice/crio-96b99b2166595d06311919f39bcf5f3bcd3dd03439156ccbcfd3b92bcdf473f0.scope\": RecentStats: unable to find data in memory cache]" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.770699 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/3.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.772257 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovnkube-controller/3.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.773941 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovn-acl-logging/0.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.774665 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovn-controller/0.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775111 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="96b99b2166595d06311919f39bcf5f3bcd3dd03439156ccbcfd3b92bcdf473f0" exitCode=0 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775143 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b" exitCode=0 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775152 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4" exitCode=0 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775162 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d" exitCode=0 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775169 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d" exitCode=0 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775177 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704" exitCode=0 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775184 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe" exitCode=143 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775191 4858 generic.go:334] "Generic (PLEG): container finished" podID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerID="dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca" exitCode=143 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775228 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"96b99b2166595d06311919f39bcf5f3bcd3dd03439156ccbcfd3b92bcdf473f0"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775261 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775272 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775283 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775290 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775299 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775306 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775334 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775343 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" event={"ID":"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1","Type":"ContainerDied","Data":"cfff7b09664ee0747c5d415bfaf39d57445ed67a89eea47f386ef37965a3205e"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775354 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfff7b09664ee0747c5d415bfaf39d57445ed67a89eea47f386ef37965a3205e" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.775372 4858 scope.go:117] "RemoveContainer" containerID="e120c8a8746a360f8fd661cf58ab426cf0096d66caa58388156eeb7272acd7a1" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.776757 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovn-acl-logging/0.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.777285 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovn-controller/0.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.777818 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.779241 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/2.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.779774 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/1.log" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.779826 4858 generic.go:334] "Generic (PLEG): container finished" podID="24656c62-314b-4c20-adf1-217d58a95f57" containerID="ba981e9a0ce9b14170b3eabfd0ccf4d14c784ab266253c7e4f5a08575832c5ac" exitCode=2 Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.779871 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerDied","Data":"ba981e9a0ce9b14170b3eabfd0ccf4d14c784ab266253c7e4f5a08575832c5ac"} Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.780381 4858 scope.go:117] "RemoveContainer" containerID="ba981e9a0ce9b14170b3eabfd0ccf4d14c784ab266253c7e4f5a08575832c5ac" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.799537 4858 scope.go:117] "RemoveContainer" containerID="3c4cf842da5644c7bbd637a5a459eaa105bcd91d27ad85dbffc2935a880d4020" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.852558 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gm5cj"] Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859167 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859205 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859225 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="nbdb" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859236 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="nbdb" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859250 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="northd" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859259 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="northd" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859273 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859282 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859299 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859327 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859337 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859345 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859355 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859363 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859377 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="sbdb" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859385 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="sbdb" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859400 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kubecfg-setup" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859408 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kubecfg-setup" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859420 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e4e001-1a9c-4669-b56a-d0d8abfe327d" containerName="oc" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859428 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e4e001-1a9c-4669-b56a-d0d8abfe327d" containerName="oc" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859441 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859453 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859464 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-acl-logging" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859474 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-acl-logging" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859487 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-node" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859496 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-node" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859609 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859626 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="sbdb" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859637 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-ovn-metrics" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859647 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="northd" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859657 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e4e001-1a9c-4669-b56a-d0d8abfe327d" containerName="oc" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859669 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859680 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="kube-rbac-proxy-node" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859689 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859697 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859707 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859716 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859728 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovn-acl-logging" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859737 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="nbdb" Mar 20 09:11:40 crc kubenswrapper[4858]: E0320 09:11:40.859868 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.859879 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" containerName="ovnkube-controller" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.864046 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974717 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-var-lib-openvswitch\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974781 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-script-lib\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974800 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-slash\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974816 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-env-overrides\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974839 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-kubelet\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974863 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-etc-openvswitch\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974883 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-systemd-units\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974909 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-ovn\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974932 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-config\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974926 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-slash" (OuterVolumeSpecName: "host-slash") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974957 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-systemd\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974977 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-node-log\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.974992 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975008 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htvhn\" (UniqueName: \"kubernetes.io/projected/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-kube-api-access-htvhn\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975027 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-var-lib-cni-networks-ovn-kubernetes\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975047 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-netd\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975067 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-bin\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975085 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-openvswitch\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975107 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-log-socket\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975126 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-netns\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975144 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-ovn-kubernetes\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975168 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovn-node-metrics-cert\") pod \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\" (UID: \"21fd7c33-ddc7-4a05-a922-472eb8ccd4e1\") " Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975257 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-run-netns\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975280 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-env-overrides\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975297 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-systemd-units\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975331 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-cni-netd\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975350 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-ovn\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975365 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-var-lib-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975387 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-node-log\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975406 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-cni-bin\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975455 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-systemd\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975475 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-kubelet\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975475 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975521 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt96z\" (UniqueName: \"kubernetes.io/projected/14b7a53e-9e83-4817-8db9-49c9d9f73be5-kube-api-access-tt96z\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975553 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovnkube-config\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975579 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975607 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovnkube-script-lib\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975629 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975653 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovn-node-metrics-cert\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975669 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-etc-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975687 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-run-ovn-kubernetes\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975703 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-log-socket\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975725 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-slash\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975770 4858 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975780 4858 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-slash\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975862 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975889 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.975910 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976152 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976205 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976224 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976243 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976266 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-log-socket" (OuterVolumeSpecName: "log-socket") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976291 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976380 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976474 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-node-log" (OuterVolumeSpecName: "node-log") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976680 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976764 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.976864 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.982798 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-kube-api-access-htvhn" (OuterVolumeSpecName: "kube-api-access-htvhn") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "kube-api-access-htvhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.984440 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:11:40 crc kubenswrapper[4858]: I0320 09:11:40.991204 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" (UID: "21fd7c33-ddc7-4a05-a922-472eb8ccd4e1"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076191 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-systemd\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076240 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-kubelet\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076267 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt96z\" (UniqueName: \"kubernetes.io/projected/14b7a53e-9e83-4817-8db9-49c9d9f73be5-kube-api-access-tt96z\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076286 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovnkube-config\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076356 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-systemd\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076393 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076419 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovnkube-script-lib\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076423 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-kubelet\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076498 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076452 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076674 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovn-node-metrics-cert\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076698 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-etc-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076716 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-run-ovn-kubernetes\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076735 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-log-socket\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076843 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-slash\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076874 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-run-netns\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076891 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-env-overrides\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076912 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-systemd-units\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076926 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-cni-netd\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076941 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-ovn\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076958 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-var-lib-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076994 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-node-log\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077011 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-cni-bin\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077059 4858 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077071 4858 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077082 4858 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077091 4858 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077102 4858 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077112 4858 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077121 4858 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-node-log\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077130 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htvhn\" (UniqueName: \"kubernetes.io/projected/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-kube-api-access-htvhn\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077140 4858 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077150 4858 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077161 4858 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077171 4858 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077180 4858 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-log-socket\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077190 4858 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077201 4858 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077212 4858 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077220 4858 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077228 4858 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077254 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-cni-bin\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077304 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-etc-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077354 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-run-ovn-kubernetes\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077382 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-log-socket\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077390 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovnkube-config\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077444 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-cni-netd\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077475 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-run-netns\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077516 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-host-slash\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077558 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-ovn\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077592 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-var-lib-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077849 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-env-overrides\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.077888 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-systemd-units\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.078057 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-node-log\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.076529 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/14b7a53e-9e83-4817-8db9-49c9d9f73be5-run-openvswitch\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.078502 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovnkube-script-lib\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.080019 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/14b7a53e-9e83-4817-8db9-49c9d9f73be5-ovn-node-metrics-cert\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.095919 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt96z\" (UniqueName: \"kubernetes.io/projected/14b7a53e-9e83-4817-8db9-49c9d9f73be5-kube-api-access-tt96z\") pod \"ovnkube-node-gm5cj\" (UID: \"14b7a53e-9e83-4817-8db9-49c9d9f73be5\") " pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.195238 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:41 crc kubenswrapper[4858]: W0320 09:11:41.213943 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14b7a53e_9e83_4817_8db9_49c9d9f73be5.slice/crio-ae45bda362a434566b407b65e87c390e82d92fbcc07e4196e422a5b289c65c49 WatchSource:0}: Error finding container ae45bda362a434566b407b65e87c390e82d92fbcc07e4196e422a5b289c65c49: Status 404 returned error can't find the container with id ae45bda362a434566b407b65e87c390e82d92fbcc07e4196e422a5b289c65c49 Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.791564 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-p2cjs_24656c62-314b-4c20-adf1-217d58a95f57/kube-multus/2.log" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.792214 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-p2cjs" event={"ID":"24656c62-314b-4c20-adf1-217d58a95f57","Type":"ContainerStarted","Data":"b42fb6d27c8e27649fca770ee29b34cb212688ebe6936caa700fc9e318b873cb"} Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.795748 4858 generic.go:334] "Generic (PLEG): container finished" podID="14b7a53e-9e83-4817-8db9-49c9d9f73be5" containerID="b245a2127064a8cad5e266ab010f36219d36a7952ee37f47b144a48534db4bb8" exitCode=0 Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.795857 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerDied","Data":"b245a2127064a8cad5e266ab010f36219d36a7952ee37f47b144a48534db4bb8"} Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.795901 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"ae45bda362a434566b407b65e87c390e82d92fbcc07e4196e422a5b289c65c49"} Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.801515 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovn-acl-logging/0.log" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.802436 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-dwpzf_21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/ovn-controller/0.log" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.803835 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-dwpzf" Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.944929 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dwpzf"] Mar 20 09:11:41 crc kubenswrapper[4858]: I0320 09:11:41.952145 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-dwpzf"] Mar 20 09:11:42 crc kubenswrapper[4858]: I0320 09:11:42.077563 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21fd7c33-ddc7-4a05-a922-472eb8ccd4e1" path="/var/lib/kubelet/pods/21fd7c33-ddc7-4a05-a922-472eb8ccd4e1/volumes" Mar 20 09:11:42 crc kubenswrapper[4858]: I0320 09:11:42.819961 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"8eb5cdd870c873a969151d90c48830be073587a52af5d920c2e36212f559afa2"} Mar 20 09:11:42 crc kubenswrapper[4858]: I0320 09:11:42.820531 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"b73797baa3d19043e81c66c344540c91366c1a0c85348af47a4ccd641eac45a9"} Mar 20 09:11:42 crc kubenswrapper[4858]: I0320 09:11:42.820549 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"89f1e3d0437ebdb9e80b3e862b5f50b31ea6bd6453910cb460d16a3fa8e2f412"} Mar 20 09:11:42 crc kubenswrapper[4858]: I0320 09:11:42.820564 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"387c72e444a98c8af7517969b458211a46e3b863a7c26a60626eab05b8634e00"} Mar 20 09:11:42 crc kubenswrapper[4858]: I0320 09:11:42.820578 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"6f5173d0d66fff0c8868409a577a08d7036b2c5bee7414642e8481718e7b165c"} Mar 20 09:11:42 crc kubenswrapper[4858]: I0320 09:11:42.820593 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"e3897f2c5cd902776c57c7af62d510b5338523b27bb1e816cf71be0bffbc048a"} Mar 20 09:11:45 crc kubenswrapper[4858]: I0320 09:11:45.847421 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"b5dabca4157ff84f8f0a3429ffc410b7d1292570a7d88a9e544b71e906134d4e"} Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.268977 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-lf2zs"] Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.270279 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.272537 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.274227 4858 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-jbz7b" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.274537 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.274542 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.364210 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7fc0136b-8b02-47f9-943d-28899c888dd9-node-mnt\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.364273 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7fc0136b-8b02-47f9-943d-28899c888dd9-crc-storage\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.364584 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6mdx\" (UniqueName: \"kubernetes.io/projected/7fc0136b-8b02-47f9-943d-28899c888dd9-kube-api-access-s6mdx\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.467153 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6mdx\" (UniqueName: \"kubernetes.io/projected/7fc0136b-8b02-47f9-943d-28899c888dd9-kube-api-access-s6mdx\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.467254 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7fc0136b-8b02-47f9-943d-28899c888dd9-node-mnt\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.467308 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7fc0136b-8b02-47f9-943d-28899c888dd9-crc-storage\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.467734 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7fc0136b-8b02-47f9-943d-28899c888dd9-node-mnt\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.468391 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7fc0136b-8b02-47f9-943d-28899c888dd9-crc-storage\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.495102 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6mdx\" (UniqueName: \"kubernetes.io/projected/7fc0136b-8b02-47f9-943d-28899c888dd9-kube-api-access-s6mdx\") pod \"crc-storage-crc-lf2zs\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: I0320 09:11:46.591455 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: E0320 09:11:46.632898 4858 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(105d5410ac0d94ed97a67893f30f866665f830d9213ca3a6d0fcf61775aee26e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:11:46 crc kubenswrapper[4858]: E0320 09:11:46.633037 4858 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(105d5410ac0d94ed97a67893f30f866665f830d9213ca3a6d0fcf61775aee26e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: E0320 09:11:46.633077 4858 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(105d5410ac0d94ed97a67893f30f866665f830d9213ca3a6d0fcf61775aee26e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:46 crc kubenswrapper[4858]: E0320 09:11:46.633138 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-lf2zs_crc-storage(7fc0136b-8b02-47f9-943d-28899c888dd9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-lf2zs_crc-storage(7fc0136b-8b02-47f9-943d-28899c888dd9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(105d5410ac0d94ed97a67893f30f866665f830d9213ca3a6d0fcf61775aee26e): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-lf2zs" podUID="7fc0136b-8b02-47f9-943d-28899c888dd9" Mar 20 09:11:48 crc kubenswrapper[4858]: I0320 09:11:48.874806 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" event={"ID":"14b7a53e-9e83-4817-8db9-49c9d9f73be5","Type":"ContainerStarted","Data":"d6eea87d62e4ac1bf9690e020865c9b876cfb2e3fefd92a9d481ff953ea828b6"} Mar 20 09:11:48 crc kubenswrapper[4858]: I0320 09:11:48.875194 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:48 crc kubenswrapper[4858]: I0320 09:11:48.875389 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:48 crc kubenswrapper[4858]: I0320 09:11:48.875423 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:48 crc kubenswrapper[4858]: I0320 09:11:48.908411 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:48 crc kubenswrapper[4858]: I0320 09:11:48.910575 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:11:48 crc kubenswrapper[4858]: I0320 09:11:48.914673 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" podStartSLOduration=8.914652622 podStartE2EDuration="8.914652622s" podCreationTimestamp="2026-03-20 09:11:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:11:48.914629132 +0000 UTC m=+890.235047369" watchObservedRunningTime="2026-03-20 09:11:48.914652622 +0000 UTC m=+890.235070829" Mar 20 09:11:49 crc kubenswrapper[4858]: I0320 09:11:49.522154 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-lf2zs"] Mar 20 09:11:49 crc kubenswrapper[4858]: I0320 09:11:49.522351 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:49 crc kubenswrapper[4858]: I0320 09:11:49.522938 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:49 crc kubenswrapper[4858]: E0320 09:11:49.585592 4858 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(1f97ad63d5f83a1dd56c8138fb3de71d7772a4f9801869739ec2127c417e0aba): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 20 09:11:49 crc kubenswrapper[4858]: E0320 09:11:49.585767 4858 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(1f97ad63d5f83a1dd56c8138fb3de71d7772a4f9801869739ec2127c417e0aba): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:49 crc kubenswrapper[4858]: E0320 09:11:49.585805 4858 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(1f97ad63d5f83a1dd56c8138fb3de71d7772a4f9801869739ec2127c417e0aba): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:11:49 crc kubenswrapper[4858]: E0320 09:11:49.585958 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-lf2zs_crc-storage(7fc0136b-8b02-47f9-943d-28899c888dd9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-lf2zs_crc-storage(7fc0136b-8b02-47f9-943d-28899c888dd9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-lf2zs_crc-storage_7fc0136b-8b02-47f9-943d-28899c888dd9_0(1f97ad63d5f83a1dd56c8138fb3de71d7772a4f9801869739ec2127c417e0aba): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-lf2zs" podUID="7fc0136b-8b02-47f9-943d-28899c888dd9" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.135177 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566632-v974z"] Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.137158 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-v974z" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.140526 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.140624 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.140828 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.145079 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-v974z"] Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.186473 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8vgs\" (UniqueName: \"kubernetes.io/projected/116c8acc-4ac8-4989-b2f2-70ff2f159f0f-kube-api-access-z8vgs\") pod \"auto-csr-approver-29566632-v974z\" (UID: \"116c8acc-4ac8-4989-b2f2-70ff2f159f0f\") " pod="openshift-infra/auto-csr-approver-29566632-v974z" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.288181 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8vgs\" (UniqueName: \"kubernetes.io/projected/116c8acc-4ac8-4989-b2f2-70ff2f159f0f-kube-api-access-z8vgs\") pod \"auto-csr-approver-29566632-v974z\" (UID: \"116c8acc-4ac8-4989-b2f2-70ff2f159f0f\") " pod="openshift-infra/auto-csr-approver-29566632-v974z" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.309211 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8vgs\" (UniqueName: \"kubernetes.io/projected/116c8acc-4ac8-4989-b2f2-70ff2f159f0f-kube-api-access-z8vgs\") pod \"auto-csr-approver-29566632-v974z\" (UID: \"116c8acc-4ac8-4989-b2f2-70ff2f159f0f\") " pod="openshift-infra/auto-csr-approver-29566632-v974z" Mar 20 09:12:00 crc kubenswrapper[4858]: I0320 09:12:00.506935 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-v974z" Mar 20 09:12:01 crc kubenswrapper[4858]: I0320 09:12:01.137350 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-v974z"] Mar 20 09:12:01 crc kubenswrapper[4858]: W0320 09:12:01.143504 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod116c8acc_4ac8_4989_b2f2_70ff2f159f0f.slice/crio-d9f4e699b8568064183f91d386f3ceaf8e2bf7ab2cc9183665bcdcaa335027cd WatchSource:0}: Error finding container d9f4e699b8568064183f91d386f3ceaf8e2bf7ab2cc9183665bcdcaa335027cd: Status 404 returned error can't find the container with id d9f4e699b8568064183f91d386f3ceaf8e2bf7ab2cc9183665bcdcaa335027cd Mar 20 09:12:01 crc kubenswrapper[4858]: I0320 09:12:01.145415 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:12:01 crc kubenswrapper[4858]: I0320 09:12:01.973380 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566632-v974z" event={"ID":"116c8acc-4ac8-4989-b2f2-70ff2f159f0f","Type":"ContainerStarted","Data":"d9f4e699b8568064183f91d386f3ceaf8e2bf7ab2cc9183665bcdcaa335027cd"} Mar 20 09:12:02 crc kubenswrapper[4858]: I0320 09:12:02.982120 4858 generic.go:334] "Generic (PLEG): container finished" podID="116c8acc-4ac8-4989-b2f2-70ff2f159f0f" containerID="afb0eaaff81e64a736ed219c95e12e0066ce9b1793557e369f41ba31fefb0942" exitCode=0 Mar 20 09:12:02 crc kubenswrapper[4858]: I0320 09:12:02.982202 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566632-v974z" event={"ID":"116c8acc-4ac8-4989-b2f2-70ff2f159f0f","Type":"ContainerDied","Data":"afb0eaaff81e64a736ed219c95e12e0066ce9b1793557e369f41ba31fefb0942"} Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.069374 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.070866 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.246223 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-v974z" Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.323847 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-lf2zs"] Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.352937 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8vgs\" (UniqueName: \"kubernetes.io/projected/116c8acc-4ac8-4989-b2f2-70ff2f159f0f-kube-api-access-z8vgs\") pod \"116c8acc-4ac8-4989-b2f2-70ff2f159f0f\" (UID: \"116c8acc-4ac8-4989-b2f2-70ff2f159f0f\") " Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.359141 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/116c8acc-4ac8-4989-b2f2-70ff2f159f0f-kube-api-access-z8vgs" (OuterVolumeSpecName: "kube-api-access-z8vgs") pod "116c8acc-4ac8-4989-b2f2-70ff2f159f0f" (UID: "116c8acc-4ac8-4989-b2f2-70ff2f159f0f"). InnerVolumeSpecName "kube-api-access-z8vgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.454794 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8vgs\" (UniqueName: \"kubernetes.io/projected/116c8acc-4ac8-4989-b2f2-70ff2f159f0f-kube-api-access-z8vgs\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.994932 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566632-v974z" event={"ID":"116c8acc-4ac8-4989-b2f2-70ff2f159f0f","Type":"ContainerDied","Data":"d9f4e699b8568064183f91d386f3ceaf8e2bf7ab2cc9183665bcdcaa335027cd"} Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.995008 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9f4e699b8568064183f91d386f3ceaf8e2bf7ab2cc9183665bcdcaa335027cd" Mar 20 09:12:04 crc kubenswrapper[4858]: I0320 09:12:04.995085 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566632-v974z" Mar 20 09:12:05 crc kubenswrapper[4858]: I0320 09:12:05.000760 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-lf2zs" event={"ID":"7fc0136b-8b02-47f9-943d-28899c888dd9","Type":"ContainerStarted","Data":"4ec479b383d874fcb9e2036987fc6e198e70af86e3e6edef59c90600b340fc3b"} Mar 20 09:12:05 crc kubenswrapper[4858]: I0320 09:12:05.315615 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-s979k"] Mar 20 09:12:05 crc kubenswrapper[4858]: I0320 09:12:05.318715 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566626-s979k"] Mar 20 09:12:06 crc kubenswrapper[4858]: I0320 09:12:06.089873 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8b8a11e-6e3c-4d93-9d52-46fe767e7b07" path="/var/lib/kubelet/pods/a8b8a11e-6e3c-4d93-9d52-46fe767e7b07/volumes" Mar 20 09:12:07 crc kubenswrapper[4858]: I0320 09:12:07.016097 4858 generic.go:334] "Generic (PLEG): container finished" podID="7fc0136b-8b02-47f9-943d-28899c888dd9" containerID="97641b666fd9d9b4d674e21c35b3d7d64a6309801317d69f59f3910f26a7c801" exitCode=0 Mar 20 09:12:07 crc kubenswrapper[4858]: I0320 09:12:07.016245 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-lf2zs" event={"ID":"7fc0136b-8b02-47f9-943d-28899c888dd9","Type":"ContainerDied","Data":"97641b666fd9d9b4d674e21c35b3d7d64a6309801317d69f59f3910f26a7c801"} Mar 20 09:12:07 crc kubenswrapper[4858]: I0320 09:12:07.890061 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:12:07 crc kubenswrapper[4858]: I0320 09:12:07.890159 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.320148 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.463934 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7fc0136b-8b02-47f9-943d-28899c888dd9-crc-storage\") pod \"7fc0136b-8b02-47f9-943d-28899c888dd9\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.464028 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7fc0136b-8b02-47f9-943d-28899c888dd9-node-mnt\") pod \"7fc0136b-8b02-47f9-943d-28899c888dd9\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.464140 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6mdx\" (UniqueName: \"kubernetes.io/projected/7fc0136b-8b02-47f9-943d-28899c888dd9-kube-api-access-s6mdx\") pod \"7fc0136b-8b02-47f9-943d-28899c888dd9\" (UID: \"7fc0136b-8b02-47f9-943d-28899c888dd9\") " Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.464227 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7fc0136b-8b02-47f9-943d-28899c888dd9-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "7fc0136b-8b02-47f9-943d-28899c888dd9" (UID: "7fc0136b-8b02-47f9-943d-28899c888dd9"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.464460 4858 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/7fc0136b-8b02-47f9-943d-28899c888dd9-node-mnt\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.470822 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc0136b-8b02-47f9-943d-28899c888dd9-kube-api-access-s6mdx" (OuterVolumeSpecName: "kube-api-access-s6mdx") pod "7fc0136b-8b02-47f9-943d-28899c888dd9" (UID: "7fc0136b-8b02-47f9-943d-28899c888dd9"). InnerVolumeSpecName "kube-api-access-s6mdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.480601 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fc0136b-8b02-47f9-943d-28899c888dd9-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "7fc0136b-8b02-47f9-943d-28899c888dd9" (UID: "7fc0136b-8b02-47f9-943d-28899c888dd9"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.565375 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6mdx\" (UniqueName: \"kubernetes.io/projected/7fc0136b-8b02-47f9-943d-28899c888dd9-kube-api-access-s6mdx\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:08 crc kubenswrapper[4858]: I0320 09:12:08.565423 4858 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/7fc0136b-8b02-47f9-943d-28899c888dd9-crc-storage\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:09 crc kubenswrapper[4858]: I0320 09:12:09.030113 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-lf2zs" event={"ID":"7fc0136b-8b02-47f9-943d-28899c888dd9","Type":"ContainerDied","Data":"4ec479b383d874fcb9e2036987fc6e198e70af86e3e6edef59c90600b340fc3b"} Mar 20 09:12:09 crc kubenswrapper[4858]: I0320 09:12:09.030180 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ec479b383d874fcb9e2036987fc6e198e70af86e3e6edef59c90600b340fc3b" Mar 20 09:12:09 crc kubenswrapper[4858]: I0320 09:12:09.030217 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-lf2zs" Mar 20 09:12:11 crc kubenswrapper[4858]: I0320 09:12:11.221955 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gm5cj" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.743974 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg"] Mar 20 09:12:15 crc kubenswrapper[4858]: E0320 09:12:15.744769 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc0136b-8b02-47f9-943d-28899c888dd9" containerName="storage" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.744784 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc0136b-8b02-47f9-943d-28899c888dd9" containerName="storage" Mar 20 09:12:15 crc kubenswrapper[4858]: E0320 09:12:15.744804 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="116c8acc-4ac8-4989-b2f2-70ff2f159f0f" containerName="oc" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.744810 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="116c8acc-4ac8-4989-b2f2-70ff2f159f0f" containerName="oc" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.744949 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="116c8acc-4ac8-4989-b2f2-70ff2f159f0f" containerName="oc" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.744964 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc0136b-8b02-47f9-943d-28899c888dd9" containerName="storage" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.746099 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.751025 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.756155 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg"] Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.772745 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.772793 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.772853 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g6s6\" (UniqueName: \"kubernetes.io/projected/20a682c7-79f0-4bf3-8574-4beee0f0415f-kube-api-access-9g6s6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.874016 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.874077 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.874112 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g6s6\" (UniqueName: \"kubernetes.io/projected/20a682c7-79f0-4bf3-8574-4beee0f0415f-kube-api-access-9g6s6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.874754 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.874924 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:15 crc kubenswrapper[4858]: I0320 09:12:15.895167 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g6s6\" (UniqueName: \"kubernetes.io/projected/20a682c7-79f0-4bf3-8574-4beee0f0415f-kube-api-access-9g6s6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:16 crc kubenswrapper[4858]: I0320 09:12:16.071558 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:16 crc kubenswrapper[4858]: I0320 09:12:16.294016 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg"] Mar 20 09:12:17 crc kubenswrapper[4858]: I0320 09:12:17.089334 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" event={"ID":"20a682c7-79f0-4bf3-8574-4beee0f0415f","Type":"ContainerStarted","Data":"d2bc296dabb664dc658f8c9c31b4d49184a5b14f8de1a562231fa24cf1eeb50f"} Mar 20 09:12:17 crc kubenswrapper[4858]: I0320 09:12:17.089812 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" event={"ID":"20a682c7-79f0-4bf3-8574-4beee0f0415f","Type":"ContainerStarted","Data":"ded8692cb5e23c274c2e447ef26665bab74fbfcaf47ca7ae273e9b95cc333129"} Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.097728 4858 generic.go:334] "Generic (PLEG): container finished" podID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerID="d2bc296dabb664dc658f8c9c31b4d49184a5b14f8de1a562231fa24cf1eeb50f" exitCode=0 Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.097847 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" event={"ID":"20a682c7-79f0-4bf3-8574-4beee0f0415f","Type":"ContainerDied","Data":"d2bc296dabb664dc658f8c9c31b4d49184a5b14f8de1a562231fa24cf1eeb50f"} Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.098908 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q5qxj"] Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.100425 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.112249 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5qxj"] Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.209084 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-utilities\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.209210 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-catalog-content\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.209340 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xfpz\" (UniqueName: \"kubernetes.io/projected/1e822ee2-d232-4b48-ae21-06bf0683afea-kube-api-access-2xfpz\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.310558 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-catalog-content\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.310655 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xfpz\" (UniqueName: \"kubernetes.io/projected/1e822ee2-d232-4b48-ae21-06bf0683afea-kube-api-access-2xfpz\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.310687 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-utilities\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.311228 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-utilities\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.311518 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-catalog-content\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.342477 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xfpz\" (UniqueName: \"kubernetes.io/projected/1e822ee2-d232-4b48-ae21-06bf0683afea-kube-api-access-2xfpz\") pod \"redhat-operators-q5qxj\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.427391 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:18 crc kubenswrapper[4858]: I0320 09:12:18.643109 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q5qxj"] Mar 20 09:12:18 crc kubenswrapper[4858]: W0320 09:12:18.649971 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e822ee2_d232_4b48_ae21_06bf0683afea.slice/crio-022362ddc93c78ac2a5af1c8f849d046da884f00c5997c45cf16aec3c17fb4be WatchSource:0}: Error finding container 022362ddc93c78ac2a5af1c8f849d046da884f00c5997c45cf16aec3c17fb4be: Status 404 returned error can't find the container with id 022362ddc93c78ac2a5af1c8f849d046da884f00c5997c45cf16aec3c17fb4be Mar 20 09:12:19 crc kubenswrapper[4858]: I0320 09:12:19.105749 4858 generic.go:334] "Generic (PLEG): container finished" podID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerID="5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b" exitCode=0 Mar 20 09:12:19 crc kubenswrapper[4858]: I0320 09:12:19.105831 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5qxj" event={"ID":"1e822ee2-d232-4b48-ae21-06bf0683afea","Type":"ContainerDied","Data":"5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b"} Mar 20 09:12:19 crc kubenswrapper[4858]: I0320 09:12:19.106266 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5qxj" event={"ID":"1e822ee2-d232-4b48-ae21-06bf0683afea","Type":"ContainerStarted","Data":"022362ddc93c78ac2a5af1c8f849d046da884f00c5997c45cf16aec3c17fb4be"} Mar 20 09:12:20 crc kubenswrapper[4858]: I0320 09:12:20.116965 4858 generic.go:334] "Generic (PLEG): container finished" podID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerID="1e548d7153feff938ebb737e54acc93b587fc86020e8706b023aa403fd9dd23d" exitCode=0 Mar 20 09:12:20 crc kubenswrapper[4858]: I0320 09:12:20.117055 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" event={"ID":"20a682c7-79f0-4bf3-8574-4beee0f0415f","Type":"ContainerDied","Data":"1e548d7153feff938ebb737e54acc93b587fc86020e8706b023aa403fd9dd23d"} Mar 20 09:12:21 crc kubenswrapper[4858]: I0320 09:12:21.129140 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5qxj" event={"ID":"1e822ee2-d232-4b48-ae21-06bf0683afea","Type":"ContainerStarted","Data":"b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32"} Mar 20 09:12:21 crc kubenswrapper[4858]: I0320 09:12:21.131992 4858 generic.go:334] "Generic (PLEG): container finished" podID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerID="7934d06d44147826b8b8bd2d322c5f706cbfe063c865f528ceca575dfb3ce62e" exitCode=0 Mar 20 09:12:21 crc kubenswrapper[4858]: I0320 09:12:21.132037 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" event={"ID":"20a682c7-79f0-4bf3-8574-4beee0f0415f","Type":"ContainerDied","Data":"7934d06d44147826b8b8bd2d322c5f706cbfe063c865f528ceca575dfb3ce62e"} Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.142528 4858 generic.go:334] "Generic (PLEG): container finished" podID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerID="b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32" exitCode=0 Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.142617 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5qxj" event={"ID":"1e822ee2-d232-4b48-ae21-06bf0683afea","Type":"ContainerDied","Data":"b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32"} Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.399330 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.492654 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g6s6\" (UniqueName: \"kubernetes.io/projected/20a682c7-79f0-4bf3-8574-4beee0f0415f-kube-api-access-9g6s6\") pod \"20a682c7-79f0-4bf3-8574-4beee0f0415f\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.492740 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-bundle\") pod \"20a682c7-79f0-4bf3-8574-4beee0f0415f\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.492776 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-util\") pod \"20a682c7-79f0-4bf3-8574-4beee0f0415f\" (UID: \"20a682c7-79f0-4bf3-8574-4beee0f0415f\") " Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.493499 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-bundle" (OuterVolumeSpecName: "bundle") pod "20a682c7-79f0-4bf3-8574-4beee0f0415f" (UID: "20a682c7-79f0-4bf3-8574-4beee0f0415f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.499151 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a682c7-79f0-4bf3-8574-4beee0f0415f-kube-api-access-9g6s6" (OuterVolumeSpecName: "kube-api-access-9g6s6") pod "20a682c7-79f0-4bf3-8574-4beee0f0415f" (UID: "20a682c7-79f0-4bf3-8574-4beee0f0415f"). InnerVolumeSpecName "kube-api-access-9g6s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.504117 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-util" (OuterVolumeSpecName: "util") pod "20a682c7-79f0-4bf3-8574-4beee0f0415f" (UID: "20a682c7-79f0-4bf3-8574-4beee0f0415f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.600279 4858 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-util\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.600405 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g6s6\" (UniqueName: \"kubernetes.io/projected/20a682c7-79f0-4bf3-8574-4beee0f0415f-kube-api-access-9g6s6\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:22 crc kubenswrapper[4858]: I0320 09:12:22.600429 4858 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/20a682c7-79f0-4bf3-8574-4beee0f0415f-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:23 crc kubenswrapper[4858]: I0320 09:12:23.151926 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" event={"ID":"20a682c7-79f0-4bf3-8574-4beee0f0415f","Type":"ContainerDied","Data":"ded8692cb5e23c274c2e447ef26665bab74fbfcaf47ca7ae273e9b95cc333129"} Mar 20 09:12:23 crc kubenswrapper[4858]: I0320 09:12:23.151990 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ded8692cb5e23c274c2e447ef26665bab74fbfcaf47ca7ae273e9b95cc333129" Mar 20 09:12:23 crc kubenswrapper[4858]: I0320 09:12:23.152028 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg" Mar 20 09:12:23 crc kubenswrapper[4858]: I0320 09:12:23.154718 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5qxj" event={"ID":"1e822ee2-d232-4b48-ae21-06bf0683afea","Type":"ContainerStarted","Data":"93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89"} Mar 20 09:12:23 crc kubenswrapper[4858]: I0320 09:12:23.179025 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q5qxj" podStartSLOduration=1.728500131 podStartE2EDuration="5.17900376s" podCreationTimestamp="2026-03-20 09:12:18 +0000 UTC" firstStartedPulling="2026-03-20 09:12:19.107869326 +0000 UTC m=+920.428287523" lastFinishedPulling="2026-03-20 09:12:22.558372955 +0000 UTC m=+923.878791152" observedRunningTime="2026-03-20 09:12:23.178353101 +0000 UTC m=+924.498771308" watchObservedRunningTime="2026-03-20 09:12:23.17900376 +0000 UTC m=+924.499421977" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.398067 4858 scope.go:117] "RemoveContainer" containerID="69cbdb6f91b1e4049b43eda8467fe6cb2b38dcccbf864b4a8f46b4b569a32704" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.416175 4858 scope.go:117] "RemoveContainer" containerID="d28258aa146ba8433b676eea707c630775fc1c64f0302caa9b3e1d67106dda2b" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.431444 4858 scope.go:117] "RemoveContainer" containerID="96b99b2166595d06311919f39bcf5f3bcd3dd03439156ccbcfd3b92bcdf473f0" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.447450 4858 scope.go:117] "RemoveContainer" containerID="1c4b5289f1779989ca82b5eeca4aeaebf2a62ae62c406295be9f3b6e5dabeb23" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.483975 4858 scope.go:117] "RemoveContainer" containerID="dbc7ef2afdf9ac0008a621af9dc479cbc7fd8be970ce09b889e530c60bed6bca" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.508169 4858 scope.go:117] "RemoveContainer" containerID="94eb86fad11e9543e653b9ae48bc941ea0c6b6ae9c738127a581f89c861369f4" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.528577 4858 scope.go:117] "RemoveContainer" containerID="cae8258de46b270aab6f7bdcc97846408fd506261ed523b182d0b3dd4f5a4a5d" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.546250 4858 scope.go:117] "RemoveContainer" containerID="5f05efce1c45c538406959f0db37e32248ba4c21d6bf69c32afdffa55c86ddfe" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.562273 4858 scope.go:117] "RemoveContainer" containerID="892423d367444c14e84b1f1d313a0749e1730e82a07296ab286339d6a7f6dc5d" Mar 20 09:12:25 crc kubenswrapper[4858]: I0320 09:12:25.579579 4858 scope.go:117] "RemoveContainer" containerID="275d99bd18e00fc45c0ec125772638720816a0f23dcb83a7e7adfd32b1aff12a" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.023549 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-q94l9"] Mar 20 09:12:26 crc kubenswrapper[4858]: E0320 09:12:26.023868 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerName="extract" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.023887 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerName="extract" Mar 20 09:12:26 crc kubenswrapper[4858]: E0320 09:12:26.023900 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerName="pull" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.023910 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerName="pull" Mar 20 09:12:26 crc kubenswrapper[4858]: E0320 09:12:26.023922 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerName="util" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.023930 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerName="util" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.024065 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a682c7-79f0-4bf3-8574-4beee0f0415f" containerName="extract" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.024599 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.030428 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-dgm8p" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.030894 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.031075 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.036898 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-q94l9"] Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.050467 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpm4m\" (UniqueName: \"kubernetes.io/projected/2105dfd4-78ef-4fd6-a179-02ad553bef8f-kube-api-access-fpm4m\") pod \"nmstate-operator-796d4cfff4-q94l9\" (UID: \"2105dfd4-78ef-4fd6-a179-02ad553bef8f\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.151758 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpm4m\" (UniqueName: \"kubernetes.io/projected/2105dfd4-78ef-4fd6-a179-02ad553bef8f-kube-api-access-fpm4m\") pod \"nmstate-operator-796d4cfff4-q94l9\" (UID: \"2105dfd4-78ef-4fd6-a179-02ad553bef8f\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.173098 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpm4m\" (UniqueName: \"kubernetes.io/projected/2105dfd4-78ef-4fd6-a179-02ad553bef8f-kube-api-access-fpm4m\") pod \"nmstate-operator-796d4cfff4-q94l9\" (UID: \"2105dfd4-78ef-4fd6-a179-02ad553bef8f\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.343962 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" Mar 20 09:12:26 crc kubenswrapper[4858]: I0320 09:12:26.825382 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-q94l9"] Mar 20 09:12:27 crc kubenswrapper[4858]: I0320 09:12:27.184828 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" event={"ID":"2105dfd4-78ef-4fd6-a179-02ad553bef8f","Type":"ContainerStarted","Data":"d1ac8ccd422b47c723facc034341e83bb4963bf13c48116d95c4f96956f59b6b"} Mar 20 09:12:28 crc kubenswrapper[4858]: I0320 09:12:28.427843 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:28 crc kubenswrapper[4858]: I0320 09:12:28.427938 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:29 crc kubenswrapper[4858]: I0320 09:12:29.475997 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q5qxj" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="registry-server" probeResult="failure" output=< Mar 20 09:12:29 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:12:29 crc kubenswrapper[4858]: > Mar 20 09:12:30 crc kubenswrapper[4858]: I0320 09:12:30.204250 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" event={"ID":"2105dfd4-78ef-4fd6-a179-02ad553bef8f","Type":"ContainerStarted","Data":"32383ec00de6e92b10373adad4cd7407cf087a82bc1b05526a748d15530babc8"} Mar 20 09:12:30 crc kubenswrapper[4858]: I0320 09:12:30.224411 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-q94l9" podStartSLOduration=1.8077880039999998 podStartE2EDuration="4.224382095s" podCreationTimestamp="2026-03-20 09:12:26 +0000 UTC" firstStartedPulling="2026-03-20 09:12:26.833960328 +0000 UTC m=+928.154378525" lastFinishedPulling="2026-03-20 09:12:29.250554419 +0000 UTC m=+930.570972616" observedRunningTime="2026-03-20 09:12:30.221502372 +0000 UTC m=+931.541920569" watchObservedRunningTime="2026-03-20 09:12:30.224382095 +0000 UTC m=+931.544800312" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.771159 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct"] Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.773855 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.775861 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bphhv" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.789006 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct"] Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.796255 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj"] Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.797021 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.805149 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.819331 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj"] Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.836421 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-b2pcs"] Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.840616 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.908269 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/49466b7e-1091-4d2f-9b3c-863941f4744d-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-hkqkj\" (UID: \"49466b7e-1091-4d2f-9b3c-863941f4744d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.908396 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm4qj\" (UniqueName: \"kubernetes.io/projected/49466b7e-1091-4d2f-9b3c-863941f4744d-kube-api-access-cm4qj\") pod \"nmstate-webhook-5f558f5558-hkqkj\" (UID: \"49466b7e-1091-4d2f-9b3c-863941f4744d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.908432 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2qf5\" (UniqueName: \"kubernetes.io/projected/f652e520-5e0b-4479-9b9b-c4abdc2c27a9-kube-api-access-q2qf5\") pod \"nmstate-metrics-9b8c8685d-6jqct\" (UID: \"f652e520-5e0b-4479-9b9b-c4abdc2c27a9\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.945905 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4"] Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.946718 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.950034 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-bzqxh" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.950187 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.959238 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4"] Mar 20 09:12:35 crc kubenswrapper[4858]: I0320 09:12:35.959831 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.009790 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz6xl\" (UniqueName: \"kubernetes.io/projected/5101bce4-d1bf-478e-82da-449ec0f98fca-kube-api-access-lz6xl\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.011060 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/49466b7e-1091-4d2f-9b3c-863941f4744d-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-hkqkj\" (UID: \"49466b7e-1091-4d2f-9b3c-863941f4744d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.012262 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-dbus-socket\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: E0320 09:12:36.012167 4858 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 20 09:12:36 crc kubenswrapper[4858]: E0320 09:12:36.014004 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49466b7e-1091-4d2f-9b3c-863941f4744d-tls-key-pair podName:49466b7e-1091-4d2f-9b3c-863941f4744d nodeName:}" failed. No retries permitted until 2026-03-20 09:12:36.513972034 +0000 UTC m=+937.834390231 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/49466b7e-1091-4d2f-9b3c-863941f4744d-tls-key-pair") pod "nmstate-webhook-5f558f5558-hkqkj" (UID: "49466b7e-1091-4d2f-9b3c-863941f4744d") : secret "openshift-nmstate-webhook" not found Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.016011 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-nmstate-lock\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.016331 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-ovs-socket\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.016500 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm4qj\" (UniqueName: \"kubernetes.io/projected/49466b7e-1091-4d2f-9b3c-863941f4744d-kube-api-access-cm4qj\") pod \"nmstate-webhook-5f558f5558-hkqkj\" (UID: \"49466b7e-1091-4d2f-9b3c-863941f4744d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.017126 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2qf5\" (UniqueName: \"kubernetes.io/projected/f652e520-5e0b-4479-9b9b-c4abdc2c27a9-kube-api-access-q2qf5\") pod \"nmstate-metrics-9b8c8685d-6jqct\" (UID: \"f652e520-5e0b-4479-9b9b-c4abdc2c27a9\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.040826 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2qf5\" (UniqueName: \"kubernetes.io/projected/f652e520-5e0b-4479-9b9b-c4abdc2c27a9-kube-api-access-q2qf5\") pod \"nmstate-metrics-9b8c8685d-6jqct\" (UID: \"f652e520-5e0b-4479-9b9b-c4abdc2c27a9\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.050275 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm4qj\" (UniqueName: \"kubernetes.io/projected/49466b7e-1091-4d2f-9b3c-863941f4744d-kube-api-access-cm4qj\") pod \"nmstate-webhook-5f558f5558-hkqkj\" (UID: \"49466b7e-1091-4d2f-9b3c-863941f4744d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.115242 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.120064 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-nmstate-lock\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.120156 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-nmstate-lock\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.120491 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-ovs-socket\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.120624 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-ovs-socket\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.120691 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7a21679b-2a77-4328-9648-14933286fb41-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.120982 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz6xl\" (UniqueName: \"kubernetes.io/projected/5101bce4-d1bf-478e-82da-449ec0f98fca-kube-api-access-lz6xl\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.121112 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvghv\" (UniqueName: \"kubernetes.io/projected/7a21679b-2a77-4328-9648-14933286fb41-kube-api-access-dvghv\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.121275 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-dbus-socket\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.121695 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a21679b-2a77-4328-9648-14933286fb41-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.121647 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5101bce4-d1bf-478e-82da-449ec0f98fca-dbus-socket\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.150399 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz6xl\" (UniqueName: \"kubernetes.io/projected/5101bce4-d1bf-478e-82da-449ec0f98fca-kube-api-access-lz6xl\") pod \"nmstate-handler-b2pcs\" (UID: \"5101bce4-d1bf-478e-82da-449ec0f98fca\") " pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.166768 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.188826 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7b7dcf6ffc-xfnb9"] Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.190207 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.214175 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b7dcf6ffc-xfnb9"] Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224454 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-oauth-serving-cert\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224514 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-trusted-ca-bundle\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224581 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvghv\" (UniqueName: \"kubernetes.io/projected/7a21679b-2a77-4328-9648-14933286fb41-kube-api-access-dvghv\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224610 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-config\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224650 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a21679b-2a77-4328-9648-14933286fb41-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224690 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-serving-cert\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224739 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7a21679b-2a77-4328-9648-14933286fb41-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224775 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfvnl\" (UniqueName: \"kubernetes.io/projected/49dc2b1b-ca0a-4e42-99a2-45503c342270-kube-api-access-vfvnl\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224801 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-service-ca\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.224826 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-oauth-config\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.226821 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/7a21679b-2a77-4328-9648-14933286fb41-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.234279 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a21679b-2a77-4328-9648-14933286fb41-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.247797 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b2pcs" event={"ID":"5101bce4-d1bf-478e-82da-449ec0f98fca","Type":"ContainerStarted","Data":"f59c4799e68cc1a11108aba0d4feea1cb42a70d79a1e0a1d15ff3175c29b2151"} Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.249466 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvghv\" (UniqueName: \"kubernetes.io/projected/7a21679b-2a77-4328-9648-14933286fb41-kube-api-access-dvghv\") pod \"nmstate-console-plugin-86f58fcf4-njbj4\" (UID: \"7a21679b-2a77-4328-9648-14933286fb41\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.267965 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.325755 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-oauth-serving-cert\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.325807 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-trusted-ca-bundle\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.325854 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-config\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.325903 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-serving-cert\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.325960 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfvnl\" (UniqueName: \"kubernetes.io/projected/49dc2b1b-ca0a-4e42-99a2-45503c342270-kube-api-access-vfvnl\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.325987 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-service-ca\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.326014 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-oauth-config\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.327229 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-trusted-ca-bundle\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.327878 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-service-ca\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.328080 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-config\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.328485 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/49dc2b1b-ca0a-4e42-99a2-45503c342270-oauth-serving-cert\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.337425 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-oauth-config\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.337889 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/49dc2b1b-ca0a-4e42-99a2-45503c342270-console-serving-cert\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.347670 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfvnl\" (UniqueName: \"kubernetes.io/projected/49dc2b1b-ca0a-4e42-99a2-45503c342270-kube-api-access-vfvnl\") pod \"console-7b7dcf6ffc-xfnb9\" (UID: \"49dc2b1b-ca0a-4e42-99a2-45503c342270\") " pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.438507 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct"] Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.495193 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4"] Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.528831 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/49466b7e-1091-4d2f-9b3c-863941f4744d-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-hkqkj\" (UID: \"49466b7e-1091-4d2f-9b3c-863941f4744d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.532589 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/49466b7e-1091-4d2f-9b3c-863941f4744d-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-hkqkj\" (UID: \"49466b7e-1091-4d2f-9b3c-863941f4744d\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.549107 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.725239 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.752482 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b7dcf6ffc-xfnb9"] Mar 20 09:12:36 crc kubenswrapper[4858]: W0320 09:12:36.758476 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49dc2b1b_ca0a_4e42_99a2_45503c342270.slice/crio-8bbdd67ed47a781fb4863a1300b56697fe41e6a0d16f6963adde8f599cebedb6 WatchSource:0}: Error finding container 8bbdd67ed47a781fb4863a1300b56697fe41e6a0d16f6963adde8f599cebedb6: Status 404 returned error can't find the container with id 8bbdd67ed47a781fb4863a1300b56697fe41e6a0d16f6963adde8f599cebedb6 Mar 20 09:12:36 crc kubenswrapper[4858]: I0320 09:12:36.921693 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj"] Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.254488 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" event={"ID":"7a21679b-2a77-4328-9648-14933286fb41","Type":"ContainerStarted","Data":"dce1bdffedffa0eafbbd474a617eeb26c65b9cdbcf07152fb59dceff6d1dbee8"} Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.256589 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b7dcf6ffc-xfnb9" event={"ID":"49dc2b1b-ca0a-4e42-99a2-45503c342270","Type":"ContainerStarted","Data":"4a4a714fd4d279f3986f8e3167ee8145998ac7e793f3c1b9de1bad1f269c4ba3"} Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.256629 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b7dcf6ffc-xfnb9" event={"ID":"49dc2b1b-ca0a-4e42-99a2-45503c342270","Type":"ContainerStarted","Data":"8bbdd67ed47a781fb4863a1300b56697fe41e6a0d16f6963adde8f599cebedb6"} Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.259995 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" event={"ID":"f652e520-5e0b-4479-9b9b-c4abdc2c27a9","Type":"ContainerStarted","Data":"bf339eacf20aa0073b83e52c4bbd3bda45c1c12a0b33acca8904ebecf98d4d03"} Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.261552 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" event={"ID":"49466b7e-1091-4d2f-9b3c-863941f4744d","Type":"ContainerStarted","Data":"ebb307c3a2361d992447800a2ab4b0f8bc5cd0c8cb46eb9bc44170d8c3fe0e1a"} Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.287904 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7b7dcf6ffc-xfnb9" podStartSLOduration=1.28787853 podStartE2EDuration="1.28787853s" podCreationTimestamp="2026-03-20 09:12:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:12:37.279746817 +0000 UTC m=+938.600165014" watchObservedRunningTime="2026-03-20 09:12:37.28787853 +0000 UTC m=+938.608296737" Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.890644 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.890729 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.890797 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.891951 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e450803b001f2a0183f7a90ae4b9b24f8c995b72aa498eab30eafb0ce280f7d"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:12:37 crc kubenswrapper[4858]: I0320 09:12:37.892032 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://2e450803b001f2a0183f7a90ae4b9b24f8c995b72aa498eab30eafb0ce280f7d" gracePeriod=600 Mar 20 09:12:38 crc kubenswrapper[4858]: I0320 09:12:38.279910 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="2e450803b001f2a0183f7a90ae4b9b24f8c995b72aa498eab30eafb0ce280f7d" exitCode=0 Mar 20 09:12:38 crc kubenswrapper[4858]: I0320 09:12:38.280001 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"2e450803b001f2a0183f7a90ae4b9b24f8c995b72aa498eab30eafb0ce280f7d"} Mar 20 09:12:38 crc kubenswrapper[4858]: I0320 09:12:38.280111 4858 scope.go:117] "RemoveContainer" containerID="3e6751ee8d22e07ea1e61646d6ebfd0426280178cdcbe58bc656568144848749" Mar 20 09:12:38 crc kubenswrapper[4858]: I0320 09:12:38.475513 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:38 crc kubenswrapper[4858]: I0320 09:12:38.518302 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:38 crc kubenswrapper[4858]: I0320 09:12:38.717758 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5qxj"] Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.299987 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-b2pcs" event={"ID":"5101bce4-d1bf-478e-82da-449ec0f98fca","Type":"ContainerStarted","Data":"3338bc9840e0205999630521258a8816e159ce498215c0ad6c934eea797c1d07"} Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.300872 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.303908 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"bef7d78bb90262eb2557357139a02f6a23b1e0a616279703c46e019d97babf79"} Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.306818 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" event={"ID":"49466b7e-1091-4d2f-9b3c-863941f4744d","Type":"ContainerStarted","Data":"708436a97b8497b917465c42c0c6d4bab506b5b4ab76f299b1115407abc5fb96"} Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.306967 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.309184 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" event={"ID":"7a21679b-2a77-4328-9648-14933286fb41","Type":"ContainerStarted","Data":"754739b85c219680611852b4eeed35ca32216344e5e877c59d486012a7b3ccf2"} Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.310417 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" event={"ID":"f652e520-5e0b-4479-9b9b-c4abdc2c27a9","Type":"ContainerStarted","Data":"3db80b4d51b89101f2bde8777ba6e6fb07fde8e22d5547d09ff5ab66e348be2c"} Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.310613 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q5qxj" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="registry-server" containerID="cri-o://93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89" gracePeriod=2 Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.326458 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-b2pcs" podStartSLOduration=2.303367923 podStartE2EDuration="5.326434283s" podCreationTimestamp="2026-03-20 09:12:35 +0000 UTC" firstStartedPulling="2026-03-20 09:12:36.208775417 +0000 UTC m=+937.529193604" lastFinishedPulling="2026-03-20 09:12:39.231841767 +0000 UTC m=+940.552259964" observedRunningTime="2026-03-20 09:12:40.320787632 +0000 UTC m=+941.641205829" watchObservedRunningTime="2026-03-20 09:12:40.326434283 +0000 UTC m=+941.646852480" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.341680 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-njbj4" podStartSLOduration=2.645072856 podStartE2EDuration="5.34165689s" podCreationTimestamp="2026-03-20 09:12:35 +0000 UTC" firstStartedPulling="2026-03-20 09:12:36.508610049 +0000 UTC m=+937.829028246" lastFinishedPulling="2026-03-20 09:12:39.205194083 +0000 UTC m=+940.525612280" observedRunningTime="2026-03-20 09:12:40.341444554 +0000 UTC m=+941.661862771" watchObservedRunningTime="2026-03-20 09:12:40.34165689 +0000 UTC m=+941.662075087" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.427211 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" podStartSLOduration=3.133881654 podStartE2EDuration="5.427186721s" podCreationTimestamp="2026-03-20 09:12:35 +0000 UTC" firstStartedPulling="2026-03-20 09:12:36.937690555 +0000 UTC m=+938.258108742" lastFinishedPulling="2026-03-20 09:12:39.230995612 +0000 UTC m=+940.551413809" observedRunningTime="2026-03-20 09:12:40.418115031 +0000 UTC m=+941.738533248" watchObservedRunningTime="2026-03-20 09:12:40.427186721 +0000 UTC m=+941.747604918" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.676666 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.801260 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-catalog-content\") pod \"1e822ee2-d232-4b48-ae21-06bf0683afea\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.801451 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-utilities\") pod \"1e822ee2-d232-4b48-ae21-06bf0683afea\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.801807 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xfpz\" (UniqueName: \"kubernetes.io/projected/1e822ee2-d232-4b48-ae21-06bf0683afea-kube-api-access-2xfpz\") pod \"1e822ee2-d232-4b48-ae21-06bf0683afea\" (UID: \"1e822ee2-d232-4b48-ae21-06bf0683afea\") " Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.802616 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-utilities" (OuterVolumeSpecName: "utilities") pod "1e822ee2-d232-4b48-ae21-06bf0683afea" (UID: "1e822ee2-d232-4b48-ae21-06bf0683afea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.809222 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e822ee2-d232-4b48-ae21-06bf0683afea-kube-api-access-2xfpz" (OuterVolumeSpecName: "kube-api-access-2xfpz") pod "1e822ee2-d232-4b48-ae21-06bf0683afea" (UID: "1e822ee2-d232-4b48-ae21-06bf0683afea"). InnerVolumeSpecName "kube-api-access-2xfpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.903519 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.903560 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xfpz\" (UniqueName: \"kubernetes.io/projected/1e822ee2-d232-4b48-ae21-06bf0683afea-kube-api-access-2xfpz\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:40 crc kubenswrapper[4858]: I0320 09:12:40.945188 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e822ee2-d232-4b48-ae21-06bf0683afea" (UID: "1e822ee2-d232-4b48-ae21-06bf0683afea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.004447 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e822ee2-d232-4b48-ae21-06bf0683afea-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.320837 4858 generic.go:334] "Generic (PLEG): container finished" podID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerID="93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89" exitCode=0 Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.320918 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q5qxj" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.320946 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5qxj" event={"ID":"1e822ee2-d232-4b48-ae21-06bf0683afea","Type":"ContainerDied","Data":"93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89"} Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.321519 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q5qxj" event={"ID":"1e822ee2-d232-4b48-ae21-06bf0683afea","Type":"ContainerDied","Data":"022362ddc93c78ac2a5af1c8f849d046da884f00c5997c45cf16aec3c17fb4be"} Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.321546 4858 scope.go:117] "RemoveContainer" containerID="93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.355931 4858 scope.go:117] "RemoveContainer" containerID="b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.365741 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q5qxj"] Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.373652 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q5qxj"] Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.389523 4858 scope.go:117] "RemoveContainer" containerID="5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.417073 4858 scope.go:117] "RemoveContainer" containerID="93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89" Mar 20 09:12:41 crc kubenswrapper[4858]: E0320 09:12:41.417570 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89\": container with ID starting with 93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89 not found: ID does not exist" containerID="93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.417633 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89"} err="failed to get container status \"93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89\": rpc error: code = NotFound desc = could not find container \"93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89\": container with ID starting with 93835688de391fa10652d0d629409a17e47447d162d36bbf14bd3df17f4deb89 not found: ID does not exist" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.417675 4858 scope.go:117] "RemoveContainer" containerID="b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32" Mar 20 09:12:41 crc kubenswrapper[4858]: E0320 09:12:41.418594 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32\": container with ID starting with b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32 not found: ID does not exist" containerID="b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.418641 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32"} err="failed to get container status \"b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32\": rpc error: code = NotFound desc = could not find container \"b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32\": container with ID starting with b45732b4081b432a32138ac5b1c019090e24304d9b8b7c0652b917b3990f6d32 not found: ID does not exist" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.418677 4858 scope.go:117] "RemoveContainer" containerID="5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b" Mar 20 09:12:41 crc kubenswrapper[4858]: E0320 09:12:41.419110 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b\": container with ID starting with 5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b not found: ID does not exist" containerID="5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b" Mar 20 09:12:41 crc kubenswrapper[4858]: I0320 09:12:41.419145 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b"} err="failed to get container status \"5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b\": rpc error: code = NotFound desc = could not find container \"5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b\": container with ID starting with 5ceeae217e1d66961fd6dcbf9f12685e6fdea67490d2d7b3ac8791a5c6359c3b not found: ID does not exist" Mar 20 09:12:42 crc kubenswrapper[4858]: I0320 09:12:42.076933 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" path="/var/lib/kubelet/pods/1e822ee2-d232-4b48-ae21-06bf0683afea/volumes" Mar 20 09:12:43 crc kubenswrapper[4858]: I0320 09:12:43.340750 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" event={"ID":"f652e520-5e0b-4479-9b9b-c4abdc2c27a9","Type":"ContainerStarted","Data":"744681e65a3f208a0bb294a2e5b626c88443df2cd916a90d827a135c081025b2"} Mar 20 09:12:43 crc kubenswrapper[4858]: I0320 09:12:43.363305 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-6jqct" podStartSLOduration=2.41104773 podStartE2EDuration="8.363274148s" podCreationTimestamp="2026-03-20 09:12:35 +0000 UTC" firstStartedPulling="2026-03-20 09:12:36.447187119 +0000 UTC m=+937.767605306" lastFinishedPulling="2026-03-20 09:12:42.399413527 +0000 UTC m=+943.719831724" observedRunningTime="2026-03-20 09:12:43.361914449 +0000 UTC m=+944.682332656" watchObservedRunningTime="2026-03-20 09:12:43.363274148 +0000 UTC m=+944.683692405" Mar 20 09:12:46 crc kubenswrapper[4858]: I0320 09:12:46.203881 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-b2pcs" Mar 20 09:12:46 crc kubenswrapper[4858]: I0320 09:12:46.549570 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:46 crc kubenswrapper[4858]: I0320 09:12:46.549648 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:46 crc kubenswrapper[4858]: I0320 09:12:46.556346 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:47 crc kubenswrapper[4858]: I0320 09:12:47.373588 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7b7dcf6ffc-xfnb9" Mar 20 09:12:47 crc kubenswrapper[4858]: I0320 09:12:47.447673 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wr84h"] Mar 20 09:12:56 crc kubenswrapper[4858]: I0320 09:12:56.731781 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-hkqkj" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.386677 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6"] Mar 20 09:13:10 crc kubenswrapper[4858]: E0320 09:13:10.387508 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="extract-utilities" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.387522 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="extract-utilities" Mar 20 09:13:10 crc kubenswrapper[4858]: E0320 09:13:10.387536 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="extract-content" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.387542 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="extract-content" Mar 20 09:13:10 crc kubenswrapper[4858]: E0320 09:13:10.387560 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="registry-server" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.387567 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="registry-server" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.387677 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e822ee2-d232-4b48-ae21-06bf0683afea" containerName="registry-server" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.388448 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.391543 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.400973 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6"] Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.489862 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.490489 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.490576 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9cpz\" (UniqueName: \"kubernetes.io/projected/59f5aa23-5099-460c-8904-d9bf6acd9958-kube-api-access-x9cpz\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.592561 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9cpz\" (UniqueName: \"kubernetes.io/projected/59f5aa23-5099-460c-8904-d9bf6acd9958-kube-api-access-x9cpz\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.593007 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.593056 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.593536 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.593572 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.616426 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9cpz\" (UniqueName: \"kubernetes.io/projected/59f5aa23-5099-460c-8904-d9bf6acd9958-kube-api-access-x9cpz\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:10 crc kubenswrapper[4858]: I0320 09:13:10.761636 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:11 crc kubenswrapper[4858]: I0320 09:13:11.195483 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6"] Mar 20 09:13:11 crc kubenswrapper[4858]: I0320 09:13:11.540500 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" event={"ID":"59f5aa23-5099-460c-8904-d9bf6acd9958","Type":"ContainerStarted","Data":"0d5a6c6aa3d88d6b3de36d108503f6024414525efba479171acf970aa6d71983"} Mar 20 09:13:11 crc kubenswrapper[4858]: I0320 09:13:11.540589 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" event={"ID":"59f5aa23-5099-460c-8904-d9bf6acd9958","Type":"ContainerStarted","Data":"ef11df517c9997721cf09bb0fab6e8322460b650912ed7a60495fa5f371ee9be"} Mar 20 09:13:12 crc kubenswrapper[4858]: I0320 09:13:12.493270 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-wr84h" podUID="91be84d3-8196-44bb-8a88-e9e6548377a1" containerName="console" containerID="cri-o://273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152" gracePeriod=15 Mar 20 09:13:12 crc kubenswrapper[4858]: I0320 09:13:12.548955 4858 generic.go:334] "Generic (PLEG): container finished" podID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerID="0d5a6c6aa3d88d6b3de36d108503f6024414525efba479171acf970aa6d71983" exitCode=0 Mar 20 09:13:12 crc kubenswrapper[4858]: I0320 09:13:12.549050 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" event={"ID":"59f5aa23-5099-460c-8904-d9bf6acd9958","Type":"ContainerDied","Data":"0d5a6c6aa3d88d6b3de36d108503f6024414525efba479171acf970aa6d71983"} Mar 20 09:13:12 crc kubenswrapper[4858]: I0320 09:13:12.965844 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wr84h_91be84d3-8196-44bb-8a88-e9e6548377a1/console/0.log" Mar 20 09:13:12 crc kubenswrapper[4858]: I0320 09:13:12.965996 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.029395 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-trusted-ca-bundle\") pod \"91be84d3-8196-44bb-8a88-e9e6548377a1\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.030921 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdvl5\" (UniqueName: \"kubernetes.io/projected/91be84d3-8196-44bb-8a88-e9e6548377a1-kube-api-access-qdvl5\") pod \"91be84d3-8196-44bb-8a88-e9e6548377a1\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.030844 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "91be84d3-8196-44bb-8a88-e9e6548377a1" (UID: "91be84d3-8196-44bb-8a88-e9e6548377a1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.030991 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-service-ca\") pod \"91be84d3-8196-44bb-8a88-e9e6548377a1\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.031872 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-service-ca" (OuterVolumeSpecName: "service-ca") pod "91be84d3-8196-44bb-8a88-e9e6548377a1" (UID: "91be84d3-8196-44bb-8a88-e9e6548377a1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.032282 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-console-config\") pod \"91be84d3-8196-44bb-8a88-e9e6548377a1\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.032732 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-oauth-config\") pod \"91be84d3-8196-44bb-8a88-e9e6548377a1\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.032789 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-serving-cert\") pod \"91be84d3-8196-44bb-8a88-e9e6548377a1\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.032803 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-console-config" (OuterVolumeSpecName: "console-config") pod "91be84d3-8196-44bb-8a88-e9e6548377a1" (UID: "91be84d3-8196-44bb-8a88-e9e6548377a1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.032911 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-oauth-serving-cert\") pod \"91be84d3-8196-44bb-8a88-e9e6548377a1\" (UID: \"91be84d3-8196-44bb-8a88-e9e6548377a1\") " Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.033302 4858 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.033453 4858 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-service-ca\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.033470 4858 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-console-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.033588 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "91be84d3-8196-44bb-8a88-e9e6548377a1" (UID: "91be84d3-8196-44bb-8a88-e9e6548377a1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.039088 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "91be84d3-8196-44bb-8a88-e9e6548377a1" (UID: "91be84d3-8196-44bb-8a88-e9e6548377a1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.039358 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91be84d3-8196-44bb-8a88-e9e6548377a1-kube-api-access-qdvl5" (OuterVolumeSpecName: "kube-api-access-qdvl5") pod "91be84d3-8196-44bb-8a88-e9e6548377a1" (UID: "91be84d3-8196-44bb-8a88-e9e6548377a1"). InnerVolumeSpecName "kube-api-access-qdvl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.041810 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "91be84d3-8196-44bb-8a88-e9e6548377a1" (UID: "91be84d3-8196-44bb-8a88-e9e6548377a1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.134939 4858 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91be84d3-8196-44bb-8a88-e9e6548377a1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.134993 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdvl5\" (UniqueName: \"kubernetes.io/projected/91be84d3-8196-44bb-8a88-e9e6548377a1-kube-api-access-qdvl5\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.135006 4858 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.135015 4858 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91be84d3-8196-44bb-8a88-e9e6548377a1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.560272 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-wr84h_91be84d3-8196-44bb-8a88-e9e6548377a1/console/0.log" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.560370 4858 generic.go:334] "Generic (PLEG): container finished" podID="91be84d3-8196-44bb-8a88-e9e6548377a1" containerID="273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152" exitCode=2 Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.560411 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wr84h" event={"ID":"91be84d3-8196-44bb-8a88-e9e6548377a1","Type":"ContainerDied","Data":"273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152"} Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.560460 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-wr84h" event={"ID":"91be84d3-8196-44bb-8a88-e9e6548377a1","Type":"ContainerDied","Data":"d1092aa8af91825451b427caff92dbdeacfc64fbef2476b11042ed5f53e4e62b"} Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.560486 4858 scope.go:117] "RemoveContainer" containerID="273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.560507 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-wr84h" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.590448 4858 scope.go:117] "RemoveContainer" containerID="273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152" Mar 20 09:13:13 crc kubenswrapper[4858]: E0320 09:13:13.595613 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152\": container with ID starting with 273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152 not found: ID does not exist" containerID="273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.595689 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152"} err="failed to get container status \"273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152\": rpc error: code = NotFound desc = could not find container \"273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152\": container with ID starting with 273dc0ee8163ed9e3ede4720476610f6df68871716f26b1b531d01cdb1681152 not found: ID does not exist" Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.613975 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-wr84h"] Mar 20 09:13:13 crc kubenswrapper[4858]: I0320 09:13:13.631095 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-wr84h"] Mar 20 09:13:14 crc kubenswrapper[4858]: I0320 09:13:14.080809 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91be84d3-8196-44bb-8a88-e9e6548377a1" path="/var/lib/kubelet/pods/91be84d3-8196-44bb-8a88-e9e6548377a1/volumes" Mar 20 09:13:14 crc kubenswrapper[4858]: I0320 09:13:14.569596 4858 generic.go:334] "Generic (PLEG): container finished" podID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerID="54f916f4fcecef4a175e21e6bd823b70dd03f1d5550920fcce3bcd3446980291" exitCode=0 Mar 20 09:13:14 crc kubenswrapper[4858]: I0320 09:13:14.570087 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" event={"ID":"59f5aa23-5099-460c-8904-d9bf6acd9958","Type":"ContainerDied","Data":"54f916f4fcecef4a175e21e6bd823b70dd03f1d5550920fcce3bcd3446980291"} Mar 20 09:13:15 crc kubenswrapper[4858]: I0320 09:13:15.584823 4858 generic.go:334] "Generic (PLEG): container finished" podID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerID="1c8d13100e980a0330899f960c9c01252f60e5cef7b8f067c5ce5232a46aa055" exitCode=0 Mar 20 09:13:15 crc kubenswrapper[4858]: I0320 09:13:15.584916 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" event={"ID":"59f5aa23-5099-460c-8904-d9bf6acd9958","Type":"ContainerDied","Data":"1c8d13100e980a0330899f960c9c01252f60e5cef7b8f067c5ce5232a46aa055"} Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.845622 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.890875 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9cpz\" (UniqueName: \"kubernetes.io/projected/59f5aa23-5099-460c-8904-d9bf6acd9958-kube-api-access-x9cpz\") pod \"59f5aa23-5099-460c-8904-d9bf6acd9958\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.891011 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-bundle\") pod \"59f5aa23-5099-460c-8904-d9bf6acd9958\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.891174 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-util\") pod \"59f5aa23-5099-460c-8904-d9bf6acd9958\" (UID: \"59f5aa23-5099-460c-8904-d9bf6acd9958\") " Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.893402 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-bundle" (OuterVolumeSpecName: "bundle") pod "59f5aa23-5099-460c-8904-d9bf6acd9958" (UID: "59f5aa23-5099-460c-8904-d9bf6acd9958"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.898577 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59f5aa23-5099-460c-8904-d9bf6acd9958-kube-api-access-x9cpz" (OuterVolumeSpecName: "kube-api-access-x9cpz") pod "59f5aa23-5099-460c-8904-d9bf6acd9958" (UID: "59f5aa23-5099-460c-8904-d9bf6acd9958"). InnerVolumeSpecName "kube-api-access-x9cpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.993187 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9cpz\" (UniqueName: \"kubernetes.io/projected/59f5aa23-5099-460c-8904-d9bf6acd9958-kube-api-access-x9cpz\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:16 crc kubenswrapper[4858]: I0320 09:13:16.993247 4858 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:17 crc kubenswrapper[4858]: I0320 09:13:17.213665 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-util" (OuterVolumeSpecName: "util") pod "59f5aa23-5099-460c-8904-d9bf6acd9958" (UID: "59f5aa23-5099-460c-8904-d9bf6acd9958"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:13:17 crc kubenswrapper[4858]: I0320 09:13:17.299303 4858 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/59f5aa23-5099-460c-8904-d9bf6acd9958-util\") on node \"crc\" DevicePath \"\"" Mar 20 09:13:17 crc kubenswrapper[4858]: I0320 09:13:17.609097 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" event={"ID":"59f5aa23-5099-460c-8904-d9bf6acd9958","Type":"ContainerDied","Data":"ef11df517c9997721cf09bb0fab6e8322460b650912ed7a60495fa5f371ee9be"} Mar 20 09:13:17 crc kubenswrapper[4858]: I0320 09:13:17.609152 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef11df517c9997721cf09bb0fab6e8322460b650912ed7a60495fa5f371ee9be" Mar 20 09:13:17 crc kubenswrapper[4858]: I0320 09:13:17.609176 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.771359 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t"] Mar 20 09:13:25 crc kubenswrapper[4858]: E0320 09:13:25.771966 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91be84d3-8196-44bb-8a88-e9e6548377a1" containerName="console" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.771979 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="91be84d3-8196-44bb-8a88-e9e6548377a1" containerName="console" Mar 20 09:13:25 crc kubenswrapper[4858]: E0320 09:13:25.771993 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerName="extract" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.771999 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerName="extract" Mar 20 09:13:25 crc kubenswrapper[4858]: E0320 09:13:25.772015 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerName="util" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.772022 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerName="util" Mar 20 09:13:25 crc kubenswrapper[4858]: E0320 09:13:25.772037 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerName="pull" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.772043 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerName="pull" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.772141 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="91be84d3-8196-44bb-8a88-e9e6548377a1" containerName="console" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.772150 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="59f5aa23-5099-460c-8904-d9bf6acd9958" containerName="extract" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.772620 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.777028 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.777253 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.777285 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.778302 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-r4ttl" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.780992 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.809276 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t"] Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.826174 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-webhook-cert\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.826268 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5gcn\" (UniqueName: \"kubernetes.io/projected/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-kube-api-access-t5gcn\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.826541 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-apiservice-cert\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.928190 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-webhook-cert\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.928279 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5gcn\" (UniqueName: \"kubernetes.io/projected/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-kube-api-access-t5gcn\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.928355 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-apiservice-cert\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.936700 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-apiservice-cert\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.937289 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-webhook-cert\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:25 crc kubenswrapper[4858]: I0320 09:13:25.946845 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5gcn\" (UniqueName: \"kubernetes.io/projected/f8fa6d2d-8e34-4685-8517-92ffa49a5dcd-kube-api-access-t5gcn\") pod \"metallb-operator-controller-manager-79f996c48b-5x54t\" (UID: \"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd\") " pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.090674 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.095885 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6"] Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.096716 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.099344 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.099562 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.099568 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-jmtwt" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.124084 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6"] Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.253653 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cc614ec0-61c6-42db-9c82-17f245a66dbb-apiservice-cert\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.254152 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nwjg\" (UniqueName: \"kubernetes.io/projected/cc614ec0-61c6-42db-9c82-17f245a66dbb-kube-api-access-6nwjg\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.254181 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cc614ec0-61c6-42db-9c82-17f245a66dbb-webhook-cert\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.355915 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nwjg\" (UniqueName: \"kubernetes.io/projected/cc614ec0-61c6-42db-9c82-17f245a66dbb-kube-api-access-6nwjg\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.355993 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cc614ec0-61c6-42db-9c82-17f245a66dbb-webhook-cert\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.356034 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cc614ec0-61c6-42db-9c82-17f245a66dbb-apiservice-cert\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.372122 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cc614ec0-61c6-42db-9c82-17f245a66dbb-webhook-cert\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.379081 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cc614ec0-61c6-42db-9c82-17f245a66dbb-apiservice-cert\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.382675 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nwjg\" (UniqueName: \"kubernetes.io/projected/cc614ec0-61c6-42db-9c82-17f245a66dbb-kube-api-access-6nwjg\") pod \"metallb-operator-webhook-server-574b899bbf-vpkw6\" (UID: \"cc614ec0-61c6-42db-9c82-17f245a66dbb\") " pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.449887 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.599632 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t"] Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.669722 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" event={"ID":"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd","Type":"ContainerStarted","Data":"49be962ad9a72a66d24abd5d0cf5704a069e2b53dabeead8e03ec0c71d224b3b"} Mar 20 09:13:26 crc kubenswrapper[4858]: I0320 09:13:26.682850 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6"] Mar 20 09:13:27 crc kubenswrapper[4858]: I0320 09:13:27.678757 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" event={"ID":"cc614ec0-61c6-42db-9c82-17f245a66dbb","Type":"ContainerStarted","Data":"3a296a154aa88c5c26c995b21bcab7525d0ec4c11e8d565c7dc78a07df42d15a"} Mar 20 09:13:32 crc kubenswrapper[4858]: I0320 09:13:32.718745 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" event={"ID":"f8fa6d2d-8e34-4685-8517-92ffa49a5dcd","Type":"ContainerStarted","Data":"e78c4ed475abc451bea1a14c349535dad31908b2fa21b20256c8dea929eccf0a"} Mar 20 09:13:32 crc kubenswrapper[4858]: I0320 09:13:32.719691 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:13:32 crc kubenswrapper[4858]: I0320 09:13:32.759721 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" podStartSLOduration=1.868058069 podStartE2EDuration="7.759694655s" podCreationTimestamp="2026-03-20 09:13:25 +0000 UTC" firstStartedPulling="2026-03-20 09:13:26.611549555 +0000 UTC m=+987.931967752" lastFinishedPulling="2026-03-20 09:13:32.503186141 +0000 UTC m=+993.823604338" observedRunningTime="2026-03-20 09:13:32.753787023 +0000 UTC m=+994.074205220" watchObservedRunningTime="2026-03-20 09:13:32.759694655 +0000 UTC m=+994.080112852" Mar 20 09:13:33 crc kubenswrapper[4858]: I0320 09:13:33.728197 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" event={"ID":"cc614ec0-61c6-42db-9c82-17f245a66dbb","Type":"ContainerStarted","Data":"53ba2b7c0a5f309f78b67a34bbdec94c8f045765ee18a64f77712cdcec04b618"} Mar 20 09:13:33 crc kubenswrapper[4858]: I0320 09:13:33.761461 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" podStartSLOduration=1.677482779 podStartE2EDuration="7.761433319s" podCreationTimestamp="2026-03-20 09:13:26 +0000 UTC" firstStartedPulling="2026-03-20 09:13:26.706132009 +0000 UTC m=+988.026550206" lastFinishedPulling="2026-03-20 09:13:32.790082549 +0000 UTC m=+994.110500746" observedRunningTime="2026-03-20 09:13:33.756295408 +0000 UTC m=+995.076713615" watchObservedRunningTime="2026-03-20 09:13:33.761433319 +0000 UTC m=+995.081851506" Mar 20 09:13:34 crc kubenswrapper[4858]: I0320 09:13:34.734496 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:46 crc kubenswrapper[4858]: I0320 09:13:46.455848 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-574b899bbf-vpkw6" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.472119 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nwtbm"] Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.474214 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.486908 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwtbm"] Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.643228 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-catalog-content\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.644089 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcl6d\" (UniqueName: \"kubernetes.io/projected/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-kube-api-access-pcl6d\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.644264 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-utilities\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.745886 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-utilities\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.746293 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-catalog-content\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.746456 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcl6d\" (UniqueName: \"kubernetes.io/projected/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-kube-api-access-pcl6d\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.747551 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-utilities\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.747718 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-catalog-content\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.771182 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcl6d\" (UniqueName: \"kubernetes.io/projected/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-kube-api-access-pcl6d\") pod \"community-operators-nwtbm\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:51 crc kubenswrapper[4858]: I0320 09:13:51.803094 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:13:52 crc kubenswrapper[4858]: I0320 09:13:52.387833 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nwtbm"] Mar 20 09:13:52 crc kubenswrapper[4858]: I0320 09:13:52.857243 4858 generic.go:334] "Generic (PLEG): container finished" podID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerID="8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11" exitCode=0 Mar 20 09:13:52 crc kubenswrapper[4858]: I0320 09:13:52.857371 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwtbm" event={"ID":"0030e2d3-fc05-4ad1-a45b-e2260383a9ad","Type":"ContainerDied","Data":"8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11"} Mar 20 09:13:52 crc kubenswrapper[4858]: I0320 09:13:52.858591 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwtbm" event={"ID":"0030e2d3-fc05-4ad1-a45b-e2260383a9ad","Type":"ContainerStarted","Data":"5f2350bef65f7712bbbefe0f8f3bd157e32a9c0aca6e8324d57af02585d8b659"} Mar 20 09:13:53 crc kubenswrapper[4858]: I0320 09:13:53.869879 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qgg8j"] Mar 20 09:13:53 crc kubenswrapper[4858]: I0320 09:13:53.870969 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:53 crc kubenswrapper[4858]: I0320 09:13:53.891936 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgg8j"] Mar 20 09:13:53 crc kubenswrapper[4858]: I0320 09:13:53.990072 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-catalog-content\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:53 crc kubenswrapper[4858]: I0320 09:13:53.990164 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfpk2\" (UniqueName: \"kubernetes.io/projected/390604e6-b428-4bd4-b5a7-4e244339148f-kube-api-access-kfpk2\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:53 crc kubenswrapper[4858]: I0320 09:13:53.990259 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-utilities\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.091763 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-catalog-content\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.091882 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfpk2\" (UniqueName: \"kubernetes.io/projected/390604e6-b428-4bd4-b5a7-4e244339148f-kube-api-access-kfpk2\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.091907 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-utilities\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.092390 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-catalog-content\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.092442 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-utilities\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.114256 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfpk2\" (UniqueName: \"kubernetes.io/projected/390604e6-b428-4bd4-b5a7-4e244339148f-kube-api-access-kfpk2\") pod \"redhat-marketplace-qgg8j\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.189548 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.494496 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgg8j"] Mar 20 09:13:54 crc kubenswrapper[4858]: W0320 09:13:54.515944 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod390604e6_b428_4bd4_b5a7_4e244339148f.slice/crio-465cd51b4cee0bf946d0e87c62bd8d4af30e83a82e8e1c8ca1d6b7280f12e591 WatchSource:0}: Error finding container 465cd51b4cee0bf946d0e87c62bd8d4af30e83a82e8e1c8ca1d6b7280f12e591: Status 404 returned error can't find the container with id 465cd51b4cee0bf946d0e87c62bd8d4af30e83a82e8e1c8ca1d6b7280f12e591 Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.884251 4858 generic.go:334] "Generic (PLEG): container finished" podID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerID="fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5" exitCode=0 Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.884351 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwtbm" event={"ID":"0030e2d3-fc05-4ad1-a45b-e2260383a9ad","Type":"ContainerDied","Data":"fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5"} Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.886470 4858 generic.go:334] "Generic (PLEG): container finished" podID="390604e6-b428-4bd4-b5a7-4e244339148f" containerID="b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf" exitCode=0 Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.886521 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgg8j" event={"ID":"390604e6-b428-4bd4-b5a7-4e244339148f","Type":"ContainerDied","Data":"b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf"} Mar 20 09:13:54 crc kubenswrapper[4858]: I0320 09:13:54.886553 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgg8j" event={"ID":"390604e6-b428-4bd4-b5a7-4e244339148f","Type":"ContainerStarted","Data":"465cd51b4cee0bf946d0e87c62bd8d4af30e83a82e8e1c8ca1d6b7280f12e591"} Mar 20 09:13:55 crc kubenswrapper[4858]: I0320 09:13:55.896353 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwtbm" event={"ID":"0030e2d3-fc05-4ad1-a45b-e2260383a9ad","Type":"ContainerStarted","Data":"b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde"} Mar 20 09:13:55 crc kubenswrapper[4858]: I0320 09:13:55.898616 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgg8j" event={"ID":"390604e6-b428-4bd4-b5a7-4e244339148f","Type":"ContainerStarted","Data":"7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160"} Mar 20 09:13:55 crc kubenswrapper[4858]: I0320 09:13:55.928090 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nwtbm" podStartSLOduration=2.280911736 podStartE2EDuration="4.928066167s" podCreationTimestamp="2026-03-20 09:13:51 +0000 UTC" firstStartedPulling="2026-03-20 09:13:52.859911408 +0000 UTC m=+1014.180329615" lastFinishedPulling="2026-03-20 09:13:55.507065849 +0000 UTC m=+1016.827484046" observedRunningTime="2026-03-20 09:13:55.92384904 +0000 UTC m=+1017.244267247" watchObservedRunningTime="2026-03-20 09:13:55.928066167 +0000 UTC m=+1017.248484364" Mar 20 09:13:56 crc kubenswrapper[4858]: I0320 09:13:56.908121 4858 generic.go:334] "Generic (PLEG): container finished" podID="390604e6-b428-4bd4-b5a7-4e244339148f" containerID="7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160" exitCode=0 Mar 20 09:13:56 crc kubenswrapper[4858]: I0320 09:13:56.908244 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgg8j" event={"ID":"390604e6-b428-4bd4-b5a7-4e244339148f","Type":"ContainerDied","Data":"7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160"} Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.665462 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z7gqf"] Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.666631 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.681443 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z7gqf"] Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.746478 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4qxf\" (UniqueName: \"kubernetes.io/projected/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-kube-api-access-w4qxf\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.746702 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-catalog-content\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.746805 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-utilities\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.848595 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-catalog-content\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.848662 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-utilities\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.848732 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4qxf\" (UniqueName: \"kubernetes.io/projected/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-kube-api-access-w4qxf\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.849264 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-catalog-content\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.849338 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-utilities\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.884732 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4qxf\" (UniqueName: \"kubernetes.io/projected/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-kube-api-access-w4qxf\") pod \"certified-operators-z7gqf\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.926431 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgg8j" event={"ID":"390604e6-b428-4bd4-b5a7-4e244339148f","Type":"ContainerStarted","Data":"a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed"} Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.953609 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qgg8j" podStartSLOduration=2.535457658 podStartE2EDuration="4.953578419s" podCreationTimestamp="2026-03-20 09:13:53 +0000 UTC" firstStartedPulling="2026-03-20 09:13:54.887901508 +0000 UTC m=+1016.208319725" lastFinishedPulling="2026-03-20 09:13:57.306022289 +0000 UTC m=+1018.626440486" observedRunningTime="2026-03-20 09:13:57.948483369 +0000 UTC m=+1019.268901576" watchObservedRunningTime="2026-03-20 09:13:57.953578419 +0000 UTC m=+1019.273996616" Mar 20 09:13:57 crc kubenswrapper[4858]: I0320 09:13:57.997960 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:13:58 crc kubenswrapper[4858]: I0320 09:13:58.317374 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z7gqf"] Mar 20 09:13:58 crc kubenswrapper[4858]: I0320 09:13:58.937412 4858 generic.go:334] "Generic (PLEG): container finished" podID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerID="f010016d70fb0cb4786309dd77245bd53410d3869e5b347c4289ca708152bfd6" exitCode=0 Mar 20 09:13:58 crc kubenswrapper[4858]: I0320 09:13:58.939040 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7gqf" event={"ID":"e0b3a9bc-2983-46ca-a0e3-1021c9f14638","Type":"ContainerDied","Data":"f010016d70fb0cb4786309dd77245bd53410d3869e5b347c4289ca708152bfd6"} Mar 20 09:13:58 crc kubenswrapper[4858]: I0320 09:13:58.939077 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7gqf" event={"ID":"e0b3a9bc-2983-46ca-a0e3-1021c9f14638","Type":"ContainerStarted","Data":"810844b757f0e3edc430c548b8aa8815d0915893e5c7779aef8a8e85abec8d54"} Mar 20 09:13:59 crc kubenswrapper[4858]: I0320 09:13:59.946807 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7gqf" event={"ID":"e0b3a9bc-2983-46ca-a0e3-1021c9f14638","Type":"ContainerStarted","Data":"4b5dd846f854616d03730e8d20db5307d3d5a5d336c33ab7aa8ed09f68659829"} Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.143630 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566634-ppj7p"] Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.144890 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-ppj7p" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.148488 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.148810 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.149927 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.161912 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-ppj7p"] Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.281512 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcgj8\" (UniqueName: \"kubernetes.io/projected/757ea99d-a6c3-4d69-a150-eb561f504a59-kube-api-access-qcgj8\") pod \"auto-csr-approver-29566634-ppj7p\" (UID: \"757ea99d-a6c3-4d69-a150-eb561f504a59\") " pod="openshift-infra/auto-csr-approver-29566634-ppj7p" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.383078 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcgj8\" (UniqueName: \"kubernetes.io/projected/757ea99d-a6c3-4d69-a150-eb561f504a59-kube-api-access-qcgj8\") pod \"auto-csr-approver-29566634-ppj7p\" (UID: \"757ea99d-a6c3-4d69-a150-eb561f504a59\") " pod="openshift-infra/auto-csr-approver-29566634-ppj7p" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.402496 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcgj8\" (UniqueName: \"kubernetes.io/projected/757ea99d-a6c3-4d69-a150-eb561f504a59-kube-api-access-qcgj8\") pod \"auto-csr-approver-29566634-ppj7p\" (UID: \"757ea99d-a6c3-4d69-a150-eb561f504a59\") " pod="openshift-infra/auto-csr-approver-29566634-ppj7p" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.465922 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-ppj7p" Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.733729 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-ppj7p"] Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.955581 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566634-ppj7p" event={"ID":"757ea99d-a6c3-4d69-a150-eb561f504a59","Type":"ContainerStarted","Data":"b83eda93233897c8d5c72a076f606a521493490d7298f447d1a3836a75a9e463"} Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.958104 4858 generic.go:334] "Generic (PLEG): container finished" podID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerID="4b5dd846f854616d03730e8d20db5307d3d5a5d336c33ab7aa8ed09f68659829" exitCode=0 Mar 20 09:14:00 crc kubenswrapper[4858]: I0320 09:14:00.958187 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7gqf" event={"ID":"e0b3a9bc-2983-46ca-a0e3-1021c9f14638","Type":"ContainerDied","Data":"4b5dd846f854616d03730e8d20db5307d3d5a5d336c33ab7aa8ed09f68659829"} Mar 20 09:14:01 crc kubenswrapper[4858]: I0320 09:14:01.803427 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:14:01 crc kubenswrapper[4858]: I0320 09:14:01.803912 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:14:01 crc kubenswrapper[4858]: I0320 09:14:01.851032 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:14:01 crc kubenswrapper[4858]: I0320 09:14:01.969206 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7gqf" event={"ID":"e0b3a9bc-2983-46ca-a0e3-1021c9f14638","Type":"ContainerStarted","Data":"25e59dbc6628661716c486ca57aeccf7acc1d98c666c4df23aaa29fe13f06772"} Mar 20 09:14:02 crc kubenswrapper[4858]: I0320 09:14:02.033606 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:14:02 crc kubenswrapper[4858]: I0320 09:14:02.068413 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z7gqf" podStartSLOduration=2.43756707 podStartE2EDuration="5.068385993s" podCreationTimestamp="2026-03-20 09:13:57 +0000 UTC" firstStartedPulling="2026-03-20 09:13:58.939922331 +0000 UTC m=+1020.260340528" lastFinishedPulling="2026-03-20 09:14:01.570741254 +0000 UTC m=+1022.891159451" observedRunningTime="2026-03-20 09:14:01.996238144 +0000 UTC m=+1023.316656351" watchObservedRunningTime="2026-03-20 09:14:02.068385993 +0000 UTC m=+1023.388804190" Mar 20 09:14:02 crc kubenswrapper[4858]: I0320 09:14:02.977733 4858 generic.go:334] "Generic (PLEG): container finished" podID="757ea99d-a6c3-4d69-a150-eb561f504a59" containerID="43a9e6bf47bed6b01e4b2601523433ff8ed6687fde7e09bdfd880700829ad0ee" exitCode=0 Mar 20 09:14:02 crc kubenswrapper[4858]: I0320 09:14:02.977839 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566634-ppj7p" event={"ID":"757ea99d-a6c3-4d69-a150-eb561f504a59","Type":"ContainerDied","Data":"43a9e6bf47bed6b01e4b2601523433ff8ed6687fde7e09bdfd880700829ad0ee"} Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.191188 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.191632 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.244155 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.259352 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwtbm"] Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.259654 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nwtbm" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="registry-server" containerID="cri-o://b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde" gracePeriod=2 Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.306918 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-ppj7p" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.454531 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcgj8\" (UniqueName: \"kubernetes.io/projected/757ea99d-a6c3-4d69-a150-eb561f504a59-kube-api-access-qcgj8\") pod \"757ea99d-a6c3-4d69-a150-eb561f504a59\" (UID: \"757ea99d-a6c3-4d69-a150-eb561f504a59\") " Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.489605 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/757ea99d-a6c3-4d69-a150-eb561f504a59-kube-api-access-qcgj8" (OuterVolumeSpecName: "kube-api-access-qcgj8") pod "757ea99d-a6c3-4d69-a150-eb561f504a59" (UID: "757ea99d-a6c3-4d69-a150-eb561f504a59"). InnerVolumeSpecName "kube-api-access-qcgj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.556669 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcgj8\" (UniqueName: \"kubernetes.io/projected/757ea99d-a6c3-4d69-a150-eb561f504a59-kube-api-access-qcgj8\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.674297 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.761630 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcl6d\" (UniqueName: \"kubernetes.io/projected/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-kube-api-access-pcl6d\") pod \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.761760 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-utilities\") pod \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.761831 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-catalog-content\") pod \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\" (UID: \"0030e2d3-fc05-4ad1-a45b-e2260383a9ad\") " Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.763189 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-utilities" (OuterVolumeSpecName: "utilities") pod "0030e2d3-fc05-4ad1-a45b-e2260383a9ad" (UID: "0030e2d3-fc05-4ad1-a45b-e2260383a9ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.764961 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-kube-api-access-pcl6d" (OuterVolumeSpecName: "kube-api-access-pcl6d") pod "0030e2d3-fc05-4ad1-a45b-e2260383a9ad" (UID: "0030e2d3-fc05-4ad1-a45b-e2260383a9ad"). InnerVolumeSpecName "kube-api-access-pcl6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.820084 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0030e2d3-fc05-4ad1-a45b-e2260383a9ad" (UID: "0030e2d3-fc05-4ad1-a45b-e2260383a9ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.863865 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.863916 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcl6d\" (UniqueName: \"kubernetes.io/projected/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-kube-api-access-pcl6d\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.863937 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0030e2d3-fc05-4ad1-a45b-e2260383a9ad-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.993657 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566634-ppj7p" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.993640 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566634-ppj7p" event={"ID":"757ea99d-a6c3-4d69-a150-eb561f504a59","Type":"ContainerDied","Data":"b83eda93233897c8d5c72a076f606a521493490d7298f447d1a3836a75a9e463"} Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.993818 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b83eda93233897c8d5c72a076f606a521493490d7298f447d1a3836a75a9e463" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.996420 4858 generic.go:334] "Generic (PLEG): container finished" podID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerID="b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde" exitCode=0 Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.997164 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nwtbm" Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.997192 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwtbm" event={"ID":"0030e2d3-fc05-4ad1-a45b-e2260383a9ad","Type":"ContainerDied","Data":"b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde"} Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.997226 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nwtbm" event={"ID":"0030e2d3-fc05-4ad1-a45b-e2260383a9ad","Type":"ContainerDied","Data":"5f2350bef65f7712bbbefe0f8f3bd157e32a9c0aca6e8324d57af02585d8b659"} Mar 20 09:14:04 crc kubenswrapper[4858]: I0320 09:14:04.997270 4858 scope.go:117] "RemoveContainer" containerID="b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.023073 4858 scope.go:117] "RemoveContainer" containerID="fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.046218 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nwtbm"] Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.055091 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.056237 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nwtbm"] Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.062200 4858 scope.go:117] "RemoveContainer" containerID="8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.079658 4858 scope.go:117] "RemoveContainer" containerID="b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde" Mar 20 09:14:05 crc kubenswrapper[4858]: E0320 09:14:05.082565 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde\": container with ID starting with b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde not found: ID does not exist" containerID="b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.082650 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde"} err="failed to get container status \"b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde\": rpc error: code = NotFound desc = could not find container \"b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde\": container with ID starting with b12bddfa722f169d3b1f09dd5cf9ddcbe439b923215fbdff9f1f2a1bc8064bde not found: ID does not exist" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.082685 4858 scope.go:117] "RemoveContainer" containerID="fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5" Mar 20 09:14:05 crc kubenswrapper[4858]: E0320 09:14:05.083287 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5\": container with ID starting with fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5 not found: ID does not exist" containerID="fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.083355 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5"} err="failed to get container status \"fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5\": rpc error: code = NotFound desc = could not find container \"fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5\": container with ID starting with fdbe2c42aacc01433c5f82c69c1a9b7235aef0e66326eba81ab8b9882624d4d5 not found: ID does not exist" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.083387 4858 scope.go:117] "RemoveContainer" containerID="8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11" Mar 20 09:14:05 crc kubenswrapper[4858]: E0320 09:14:05.083762 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11\": container with ID starting with 8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11 not found: ID does not exist" containerID="8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.083783 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11"} err="failed to get container status \"8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11\": rpc error: code = NotFound desc = could not find container \"8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11\": container with ID starting with 8c810f0910e996048b0ceb43ba70538a1f6b5eb746feba9fea6bcf3511657e11 not found: ID does not exist" Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.377094 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-54mc2"] Mar 20 09:14:05 crc kubenswrapper[4858]: I0320 09:14:05.380924 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566628-54mc2"] Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.078259 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" path="/var/lib/kubelet/pods/0030e2d3-fc05-4ad1-a45b-e2260383a9ad/volumes" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.078920 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f99eb15-1e3e-4ed9-932d-991b29255c03" path="/var/lib/kubelet/pods/7f99eb15-1e3e-4ed9-932d-991b29255c03/volumes" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.093643 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-79f996c48b-5x54t" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.767974 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-lwq5q"] Mar 20 09:14:06 crc kubenswrapper[4858]: E0320 09:14:06.768800 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="extract-utilities" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.768814 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="extract-utilities" Mar 20 09:14:06 crc kubenswrapper[4858]: E0320 09:14:06.768833 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="757ea99d-a6c3-4d69-a150-eb561f504a59" containerName="oc" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.768839 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="757ea99d-a6c3-4d69-a150-eb561f504a59" containerName="oc" Mar 20 09:14:06 crc kubenswrapper[4858]: E0320 09:14:06.768853 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="extract-content" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.768861 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="extract-content" Mar 20 09:14:06 crc kubenswrapper[4858]: E0320 09:14:06.768869 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="registry-server" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.768875 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="registry-server" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.768979 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="757ea99d-a6c3-4d69-a150-eb561f504a59" containerName="oc" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.768987 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="0030e2d3-fc05-4ad1-a45b-e2260383a9ad" containerName="registry-server" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.770918 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.775815 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.776355 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-mgjkh" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.780734 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.812011 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95"] Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.812888 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.814563 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.841945 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95"] Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.905727 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-startup\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906014 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62cd2367-51df-4a29-a892-cb01bbf2a98b-metrics-certs\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906184 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df73c110-c68f-4dd3-b70b-e0898869c0a6-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-2jb95\" (UID: \"df73c110-c68f-4dd3-b70b-e0898869c0a6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906253 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-sockets\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906425 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-reloader\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906527 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqxsw\" (UniqueName: \"kubernetes.io/projected/62cd2367-51df-4a29-a892-cb01bbf2a98b-kube-api-access-vqxsw\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906557 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-metrics\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906708 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-conf\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.906740 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zn4p\" (UniqueName: \"kubernetes.io/projected/df73c110-c68f-4dd3-b70b-e0898869c0a6-kube-api-access-4zn4p\") pod \"frr-k8s-webhook-server-bcc4b6f68-2jb95\" (UID: \"df73c110-c68f-4dd3-b70b-e0898869c0a6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.943864 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-h4sxs"] Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.945032 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h4sxs" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.950212 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.950547 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.950825 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.951042 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qjgpt" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.980255 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-dm6h9"] Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.981459 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:06 crc kubenswrapper[4858]: I0320 09:14:06.983860 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.002388 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-dm6h9"] Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.046751 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks87t\" (UniqueName: \"kubernetes.io/projected/b2351ddb-14a8-445f-9326-9e49d955e417-kube-api-access-ks87t\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.046798 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-conf\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.046829 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zn4p\" (UniqueName: \"kubernetes.io/projected/df73c110-c68f-4dd3-b70b-e0898869c0a6-kube-api-access-4zn4p\") pod \"frr-k8s-webhook-server-bcc4b6f68-2jb95\" (UID: \"df73c110-c68f-4dd3-b70b-e0898869c0a6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047021 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-startup\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047113 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62cd2367-51df-4a29-a892-cb01bbf2a98b-metrics-certs\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047145 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df73c110-c68f-4dd3-b70b-e0898869c0a6-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-2jb95\" (UID: \"df73c110-c68f-4dd3-b70b-e0898869c0a6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047171 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-sockets\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047250 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-conf\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047219 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b2351ddb-14a8-445f-9326-9e49d955e417-metallb-excludel2\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047381 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwfzc\" (UniqueName: \"kubernetes.io/projected/e548e45c-1cb2-48a4-bc75-679f254219e5-kube-api-access-fwfzc\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047459 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e548e45c-1cb2-48a4-bc75-679f254219e5-cert\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047513 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-reloader\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047552 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e548e45c-1cb2-48a4-bc75-679f254219e5-metrics-certs\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047584 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-metrics-certs\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047670 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqxsw\" (UniqueName: \"kubernetes.io/projected/62cd2367-51df-4a29-a892-cb01bbf2a98b-kube-api-access-vqxsw\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047705 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.047750 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-metrics\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.048014 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-startup\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.048122 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-metrics\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.048391 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-reloader\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.048925 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/62cd2367-51df-4a29-a892-cb01bbf2a98b-frr-sockets\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.055753 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/df73c110-c68f-4dd3-b70b-e0898869c0a6-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-2jb95\" (UID: \"df73c110-c68f-4dd3-b70b-e0898869c0a6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.055753 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/62cd2367-51df-4a29-a892-cb01bbf2a98b-metrics-certs\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.079080 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqxsw\" (UniqueName: \"kubernetes.io/projected/62cd2367-51df-4a29-a892-cb01bbf2a98b-kube-api-access-vqxsw\") pod \"frr-k8s-lwq5q\" (UID: \"62cd2367-51df-4a29-a892-cb01bbf2a98b\") " pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.095026 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zn4p\" (UniqueName: \"kubernetes.io/projected/df73c110-c68f-4dd3-b70b-e0898869c0a6-kube-api-access-4zn4p\") pod \"frr-k8s-webhook-server-bcc4b6f68-2jb95\" (UID: \"df73c110-c68f-4dd3-b70b-e0898869c0a6\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.095717 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.136266 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154003 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b2351ddb-14a8-445f-9326-9e49d955e417-metallb-excludel2\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154076 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwfzc\" (UniqueName: \"kubernetes.io/projected/e548e45c-1cb2-48a4-bc75-679f254219e5-kube-api-access-fwfzc\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154129 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e548e45c-1cb2-48a4-bc75-679f254219e5-cert\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154164 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-metrics-certs\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154181 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e548e45c-1cb2-48a4-bc75-679f254219e5-metrics-certs\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154251 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154393 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks87t\" (UniqueName: \"kubernetes.io/projected/b2351ddb-14a8-445f-9326-9e49d955e417-kube-api-access-ks87t\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: E0320 09:14:07.154439 4858 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Mar 20 09:14:07 crc kubenswrapper[4858]: E0320 09:14:07.154519 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-metrics-certs podName:b2351ddb-14a8-445f-9326-9e49d955e417 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:07.654494775 +0000 UTC m=+1028.974912972 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-metrics-certs") pod "speaker-h4sxs" (UID: "b2351ddb-14a8-445f-9326-9e49d955e417") : secret "speaker-certs-secret" not found Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.154827 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b2351ddb-14a8-445f-9326-9e49d955e417-metallb-excludel2\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: E0320 09:14:07.154896 4858 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 20 09:14:07 crc kubenswrapper[4858]: E0320 09:14:07.155027 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist podName:b2351ddb-14a8-445f-9326-9e49d955e417 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:07.654989879 +0000 UTC m=+1028.975408276 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist") pod "speaker-h4sxs" (UID: "b2351ddb-14a8-445f-9326-9e49d955e417") : secret "metallb-memberlist" not found Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.156966 4858 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.158107 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e548e45c-1cb2-48a4-bc75-679f254219e5-metrics-certs\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.168560 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e548e45c-1cb2-48a4-bc75-679f254219e5-cert\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.171978 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwfzc\" (UniqueName: \"kubernetes.io/projected/e548e45c-1cb2-48a4-bc75-679f254219e5-kube-api-access-fwfzc\") pod \"controller-7bb4cc7c98-dm6h9\" (UID: \"e548e45c-1cb2-48a4-bc75-679f254219e5\") " pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.176508 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks87t\" (UniqueName: \"kubernetes.io/projected/b2351ddb-14a8-445f-9326-9e49d955e417-kube-api-access-ks87t\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.301778 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.396116 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95"] Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.541187 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-dm6h9"] Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.661528 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.661646 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-metrics-certs\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: E0320 09:14:07.662201 4858 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 20 09:14:07 crc kubenswrapper[4858]: E0320 09:14:07.662343 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist podName:b2351ddb-14a8-445f-9326-9e49d955e417 nodeName:}" failed. No retries permitted until 2026-03-20 09:14:08.662294152 +0000 UTC m=+1029.982712349 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist") pod "speaker-h4sxs" (UID: "b2351ddb-14a8-445f-9326-9e49d955e417") : secret "metallb-memberlist" not found Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.670754 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-metrics-certs\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.998634 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:14:07 crc kubenswrapper[4858]: I0320 09:14:07.998711 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.056884 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.060667 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgg8j"] Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.060980 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qgg8j" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="registry-server" containerID="cri-o://a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed" gracePeriod=2 Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.067140 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-dm6h9" event={"ID":"e548e45c-1cb2-48a4-bc75-679f254219e5","Type":"ContainerStarted","Data":"75ac342bbd722de67e5bc9891c2763f2eac89e26d6b445b70534d3aeb5d10381"} Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.067193 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-dm6h9" event={"ID":"e548e45c-1cb2-48a4-bc75-679f254219e5","Type":"ContainerStarted","Data":"c9b3ff5d7c1a366236ff34539422d202db17f2463e48818aa87f5ad0f9748af6"} Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.067207 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-dm6h9" event={"ID":"e548e45c-1cb2-48a4-bc75-679f254219e5","Type":"ContainerStarted","Data":"141fe26613ea47b5e5475ac90c02b7639b802cfc95eb76779f1688d451c8df6c"} Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.067582 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.110551 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerStarted","Data":"993db882ed40d1b74d4ca7cc4ce0eb3efde55520498366a66d0e00db853c5c9a"} Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.110638 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" event={"ID":"df73c110-c68f-4dd3-b70b-e0898869c0a6","Type":"ContainerStarted","Data":"16ce3771b374be5e902ccb3dc312e34086168e77f693268921c5599fb0959b82"} Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.124515 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-dm6h9" podStartSLOduration=2.124472608 podStartE2EDuration="2.124472608s" podCreationTimestamp="2026-03-20 09:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:14:08.114576017 +0000 UTC m=+1029.434994224" watchObservedRunningTime="2026-03-20 09:14:08.124472608 +0000 UTC m=+1029.444890805" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.153407 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.481845 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.578001 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-utilities\") pod \"390604e6-b428-4bd4-b5a7-4e244339148f\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.579060 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-catalog-content\") pod \"390604e6-b428-4bd4-b5a7-4e244339148f\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.579142 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfpk2\" (UniqueName: \"kubernetes.io/projected/390604e6-b428-4bd4-b5a7-4e244339148f-kube-api-access-kfpk2\") pod \"390604e6-b428-4bd4-b5a7-4e244339148f\" (UID: \"390604e6-b428-4bd4-b5a7-4e244339148f\") " Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.578968 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-utilities" (OuterVolumeSpecName: "utilities") pod "390604e6-b428-4bd4-b5a7-4e244339148f" (UID: "390604e6-b428-4bd4-b5a7-4e244339148f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.589386 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/390604e6-b428-4bd4-b5a7-4e244339148f-kube-api-access-kfpk2" (OuterVolumeSpecName: "kube-api-access-kfpk2") pod "390604e6-b428-4bd4-b5a7-4e244339148f" (UID: "390604e6-b428-4bd4-b5a7-4e244339148f"). InnerVolumeSpecName "kube-api-access-kfpk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.607252 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "390604e6-b428-4bd4-b5a7-4e244339148f" (UID: "390604e6-b428-4bd4-b5a7-4e244339148f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.680368 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.680502 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfpk2\" (UniqueName: \"kubernetes.io/projected/390604e6-b428-4bd4-b5a7-4e244339148f-kube-api-access-kfpk2\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.680519 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.680534 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/390604e6-b428-4bd4-b5a7-4e244339148f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.685738 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b2351ddb-14a8-445f-9326-9e49d955e417-memberlist\") pod \"speaker-h4sxs\" (UID: \"b2351ddb-14a8-445f-9326-9e49d955e417\") " pod="metallb-system/speaker-h4sxs" Mar 20 09:14:08 crc kubenswrapper[4858]: I0320 09:14:08.765075 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-h4sxs" Mar 20 09:14:08 crc kubenswrapper[4858]: W0320 09:14:08.799006 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2351ddb_14a8_445f_9326_9e49d955e417.slice/crio-cd4eb5e5c9086aaf01e576bfe21f8ea2de2e8b1f5444529293df4ef4becd5371 WatchSource:0}: Error finding container cd4eb5e5c9086aaf01e576bfe21f8ea2de2e8b1f5444529293df4ef4becd5371: Status 404 returned error can't find the container with id cd4eb5e5c9086aaf01e576bfe21f8ea2de2e8b1f5444529293df4ef4becd5371 Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.117065 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h4sxs" event={"ID":"b2351ddb-14a8-445f-9326-9e49d955e417","Type":"ContainerStarted","Data":"cd4eb5e5c9086aaf01e576bfe21f8ea2de2e8b1f5444529293df4ef4becd5371"} Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.119355 4858 generic.go:334] "Generic (PLEG): container finished" podID="390604e6-b428-4bd4-b5a7-4e244339148f" containerID="a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed" exitCode=0 Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.119459 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgg8j" event={"ID":"390604e6-b428-4bd4-b5a7-4e244339148f","Type":"ContainerDied","Data":"a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed"} Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.119510 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qgg8j" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.119522 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qgg8j" event={"ID":"390604e6-b428-4bd4-b5a7-4e244339148f","Type":"ContainerDied","Data":"465cd51b4cee0bf946d0e87c62bd8d4af30e83a82e8e1c8ca1d6b7280f12e591"} Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.119545 4858 scope.go:117] "RemoveContainer" containerID="a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.145832 4858 scope.go:117] "RemoveContainer" containerID="7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.167792 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgg8j"] Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.174526 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qgg8j"] Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.185185 4858 scope.go:117] "RemoveContainer" containerID="b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.226443 4858 scope.go:117] "RemoveContainer" containerID="a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed" Mar 20 09:14:09 crc kubenswrapper[4858]: E0320 09:14:09.229235 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed\": container with ID starting with a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed not found: ID does not exist" containerID="a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.229270 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed"} err="failed to get container status \"a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed\": rpc error: code = NotFound desc = could not find container \"a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed\": container with ID starting with a302b9775718feee6a5725538340bfb75b0e56d049e9fca75012fc9bbc0b53ed not found: ID does not exist" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.229293 4858 scope.go:117] "RemoveContainer" containerID="7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160" Mar 20 09:14:09 crc kubenswrapper[4858]: E0320 09:14:09.232332 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160\": container with ID starting with 7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160 not found: ID does not exist" containerID="7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.232363 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160"} err="failed to get container status \"7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160\": rpc error: code = NotFound desc = could not find container \"7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160\": container with ID starting with 7f0fb28fac3f79eca4cc0abe5323aeb8af9f6aa647d88ccce2507b02a2aa6160 not found: ID does not exist" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.232378 4858 scope.go:117] "RemoveContainer" containerID="b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf" Mar 20 09:14:09 crc kubenswrapper[4858]: E0320 09:14:09.235624 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf\": container with ID starting with b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf not found: ID does not exist" containerID="b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf" Mar 20 09:14:09 crc kubenswrapper[4858]: I0320 09:14:09.235661 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf"} err="failed to get container status \"b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf\": rpc error: code = NotFound desc = could not find container \"b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf\": container with ID starting with b8343f925fba9533a37ccabd88ff9dd3b7db31ea26b1083fa9e2a59daffc30bf not found: ID does not exist" Mar 20 09:14:10 crc kubenswrapper[4858]: I0320 09:14:10.080063 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" path="/var/lib/kubelet/pods/390604e6-b428-4bd4-b5a7-4e244339148f/volumes" Mar 20 09:14:10 crc kubenswrapper[4858]: I0320 09:14:10.158133 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h4sxs" event={"ID":"b2351ddb-14a8-445f-9326-9e49d955e417","Type":"ContainerStarted","Data":"ffd4f88f5b8b7c1e86d7dd89aa2d5e25b1f17ef351b7d076108071384aa6a75a"} Mar 20 09:14:10 crc kubenswrapper[4858]: I0320 09:14:10.158230 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-h4sxs" event={"ID":"b2351ddb-14a8-445f-9326-9e49d955e417","Type":"ContainerStarted","Data":"1d2e85735b0b19238a0516784dc8775cc5c8724092b52c258925017e51e8f377"} Mar 20 09:14:10 crc kubenswrapper[4858]: I0320 09:14:10.159607 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-h4sxs" Mar 20 09:14:10 crc kubenswrapper[4858]: I0320 09:14:10.187728 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-h4sxs" podStartSLOduration=4.187700545 podStartE2EDuration="4.187700545s" podCreationTimestamp="2026-03-20 09:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:14:10.184804726 +0000 UTC m=+1031.505222923" watchObservedRunningTime="2026-03-20 09:14:10.187700545 +0000 UTC m=+1031.508118742" Mar 20 09:14:10 crc kubenswrapper[4858]: I0320 09:14:10.669955 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z7gqf"] Mar 20 09:14:10 crc kubenswrapper[4858]: I0320 09:14:10.670282 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z7gqf" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="registry-server" containerID="cri-o://25e59dbc6628661716c486ca57aeccf7acc1d98c666c4df23aaa29fe13f06772" gracePeriod=2 Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.191056 4858 generic.go:334] "Generic (PLEG): container finished" podID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerID="25e59dbc6628661716c486ca57aeccf7acc1d98c666c4df23aaa29fe13f06772" exitCode=0 Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.191117 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7gqf" event={"ID":"e0b3a9bc-2983-46ca-a0e3-1021c9f14638","Type":"ContainerDied","Data":"25e59dbc6628661716c486ca57aeccf7acc1d98c666c4df23aaa29fe13f06772"} Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.331640 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.439632 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-catalog-content\") pod \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.439706 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4qxf\" (UniqueName: \"kubernetes.io/projected/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-kube-api-access-w4qxf\") pod \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.439778 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-utilities\") pod \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\" (UID: \"e0b3a9bc-2983-46ca-a0e3-1021c9f14638\") " Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.442689 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-utilities" (OuterVolumeSpecName: "utilities") pod "e0b3a9bc-2983-46ca-a0e3-1021c9f14638" (UID: "e0b3a9bc-2983-46ca-a0e3-1021c9f14638"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.450817 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-kube-api-access-w4qxf" (OuterVolumeSpecName: "kube-api-access-w4qxf") pod "e0b3a9bc-2983-46ca-a0e3-1021c9f14638" (UID: "e0b3a9bc-2983-46ca-a0e3-1021c9f14638"). InnerVolumeSpecName "kube-api-access-w4qxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.505612 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0b3a9bc-2983-46ca-a0e3-1021c9f14638" (UID: "e0b3a9bc-2983-46ca-a0e3-1021c9f14638"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.541716 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.541755 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:12 crc kubenswrapper[4858]: I0320 09:14:12.541769 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4qxf\" (UniqueName: \"kubernetes.io/projected/e0b3a9bc-2983-46ca-a0e3-1021c9f14638-kube-api-access-w4qxf\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:13 crc kubenswrapper[4858]: I0320 09:14:13.204199 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z7gqf" event={"ID":"e0b3a9bc-2983-46ca-a0e3-1021c9f14638","Type":"ContainerDied","Data":"810844b757f0e3edc430c548b8aa8815d0915893e5c7779aef8a8e85abec8d54"} Mar 20 09:14:13 crc kubenswrapper[4858]: I0320 09:14:13.204285 4858 scope.go:117] "RemoveContainer" containerID="25e59dbc6628661716c486ca57aeccf7acc1d98c666c4df23aaa29fe13f06772" Mar 20 09:14:13 crc kubenswrapper[4858]: I0320 09:14:13.204405 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z7gqf" Mar 20 09:14:13 crc kubenswrapper[4858]: I0320 09:14:13.253970 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z7gqf"] Mar 20 09:14:13 crc kubenswrapper[4858]: I0320 09:14:13.259816 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z7gqf"] Mar 20 09:14:14 crc kubenswrapper[4858]: I0320 09:14:14.080450 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" path="/var/lib/kubelet/pods/e0b3a9bc-2983-46ca-a0e3-1021c9f14638/volumes" Mar 20 09:14:15 crc kubenswrapper[4858]: I0320 09:14:15.200433 4858 scope.go:117] "RemoveContainer" containerID="4b5dd846f854616d03730e8d20db5307d3d5a5d336c33ab7aa8ed09f68659829" Mar 20 09:14:15 crc kubenswrapper[4858]: I0320 09:14:15.257384 4858 scope.go:117] "RemoveContainer" containerID="f010016d70fb0cb4786309dd77245bd53410d3869e5b347c4289ca708152bfd6" Mar 20 09:14:16 crc kubenswrapper[4858]: I0320 09:14:16.225976 4858 generic.go:334] "Generic (PLEG): container finished" podID="62cd2367-51df-4a29-a892-cb01bbf2a98b" containerID="868b299177c7363235cfd96bf4395fa1eada556a54d3f6e02a55dccd850a32a0" exitCode=0 Mar 20 09:14:16 crc kubenswrapper[4858]: I0320 09:14:16.226518 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerDied","Data":"868b299177c7363235cfd96bf4395fa1eada556a54d3f6e02a55dccd850a32a0"} Mar 20 09:14:16 crc kubenswrapper[4858]: I0320 09:14:16.237197 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" event={"ID":"df73c110-c68f-4dd3-b70b-e0898869c0a6","Type":"ContainerStarted","Data":"6026e271d4fc9886e8fa89967f7d3aeb4218b22ef9fb7c3fb40e0b5a28e53dc9"} Mar 20 09:14:16 crc kubenswrapper[4858]: I0320 09:14:16.238120 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:16 crc kubenswrapper[4858]: I0320 09:14:16.300934 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" podStartSLOduration=2.403047205 podStartE2EDuration="10.300903822s" podCreationTimestamp="2026-03-20 09:14:06 +0000 UTC" firstStartedPulling="2026-03-20 09:14:07.417000485 +0000 UTC m=+1028.737418682" lastFinishedPulling="2026-03-20 09:14:15.314857082 +0000 UTC m=+1036.635275299" observedRunningTime="2026-03-20 09:14:16.294042835 +0000 UTC m=+1037.614461042" watchObservedRunningTime="2026-03-20 09:14:16.300903822 +0000 UTC m=+1037.621322019" Mar 20 09:14:17 crc kubenswrapper[4858]: I0320 09:14:17.250091 4858 generic.go:334] "Generic (PLEG): container finished" podID="62cd2367-51df-4a29-a892-cb01bbf2a98b" containerID="1b149e7076bdec8d7ca480ca4fa962932440de5e4c65a9971156e63f6ad13411" exitCode=0 Mar 20 09:14:17 crc kubenswrapper[4858]: I0320 09:14:17.250162 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerDied","Data":"1b149e7076bdec8d7ca480ca4fa962932440de5e4c65a9971156e63f6ad13411"} Mar 20 09:14:17 crc kubenswrapper[4858]: I0320 09:14:17.307162 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-dm6h9" Mar 20 09:14:18 crc kubenswrapper[4858]: I0320 09:14:18.258188 4858 generic.go:334] "Generic (PLEG): container finished" podID="62cd2367-51df-4a29-a892-cb01bbf2a98b" containerID="ae50307d889203f436a0077400a07366ca8d9d9e5a4cf69a899ac21b4d1ab225" exitCode=0 Mar 20 09:14:18 crc kubenswrapper[4858]: I0320 09:14:18.258255 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerDied","Data":"ae50307d889203f436a0077400a07366ca8d9d9e5a4cf69a899ac21b4d1ab225"} Mar 20 09:14:19 crc kubenswrapper[4858]: I0320 09:14:19.270118 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerStarted","Data":"108b81e7b445d1a762967ec2615bab0ede201570e41380fc88113be3d74c2564"} Mar 20 09:14:19 crc kubenswrapper[4858]: I0320 09:14:19.270594 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerStarted","Data":"2313687b35526c859cb0a6c3038224490c015e499f857d5abe74280f86149e64"} Mar 20 09:14:19 crc kubenswrapper[4858]: I0320 09:14:19.270608 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerStarted","Data":"d219722a82936fc4920dfe02fcfbc31ceca60c45983a3a3f5855c29377c1c249"} Mar 20 09:14:19 crc kubenswrapper[4858]: I0320 09:14:19.270748 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerStarted","Data":"193b16a7511a1277285876de71ea30aaef791b1674f2a755d94ae225dc1d22e1"} Mar 20 09:14:19 crc kubenswrapper[4858]: I0320 09:14:19.270767 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerStarted","Data":"cf494cfaa5902e1423bcba8f2d30e6690e44711352ef5dc5e647bdc91dd18270"} Mar 20 09:14:20 crc kubenswrapper[4858]: I0320 09:14:20.282162 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lwq5q" event={"ID":"62cd2367-51df-4a29-a892-cb01bbf2a98b","Type":"ContainerStarted","Data":"7cf18f6c48aa06d31df7e4ff608ac5e8f12babab95661d9908e9d0fa4d586429"} Mar 20 09:14:20 crc kubenswrapper[4858]: I0320 09:14:20.282914 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:20 crc kubenswrapper[4858]: I0320 09:14:20.322650 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-lwq5q" podStartSLOduration=6.343788818 podStartE2EDuration="14.32262882s" podCreationTimestamp="2026-03-20 09:14:06 +0000 UTC" firstStartedPulling="2026-03-20 09:14:07.313439675 +0000 UTC m=+1028.633857872" lastFinishedPulling="2026-03-20 09:14:15.292279687 +0000 UTC m=+1036.612697874" observedRunningTime="2026-03-20 09:14:20.318080992 +0000 UTC m=+1041.638499209" watchObservedRunningTime="2026-03-20 09:14:20.32262882 +0000 UTC m=+1041.643047017" Mar 20 09:14:22 crc kubenswrapper[4858]: I0320 09:14:22.096514 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:22 crc kubenswrapper[4858]: I0320 09:14:22.143652 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:25 crc kubenswrapper[4858]: I0320 09:14:25.650950 4858 scope.go:117] "RemoveContainer" containerID="e8534a5e53e6fda3045f827a8b7a484525243445c73e3763e967a16eaa46ebf0" Mar 20 09:14:27 crc kubenswrapper[4858]: I0320 09:14:27.143018 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2jb95" Mar 20 09:14:28 crc kubenswrapper[4858]: I0320 09:14:28.768939 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-h4sxs" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145080 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d"] Mar 20 09:14:30 crc kubenswrapper[4858]: E0320 09:14:30.145395 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="extract-content" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145434 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="extract-content" Mar 20 09:14:30 crc kubenswrapper[4858]: E0320 09:14:30.145455 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="extract-content" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145462 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="extract-content" Mar 20 09:14:30 crc kubenswrapper[4858]: E0320 09:14:30.145475 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="registry-server" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145483 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="registry-server" Mar 20 09:14:30 crc kubenswrapper[4858]: E0320 09:14:30.145495 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="extract-utilities" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145502 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="extract-utilities" Mar 20 09:14:30 crc kubenswrapper[4858]: E0320 09:14:30.145512 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="extract-utilities" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145519 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="extract-utilities" Mar 20 09:14:30 crc kubenswrapper[4858]: E0320 09:14:30.145538 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="registry-server" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145547 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="registry-server" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145673 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0b3a9bc-2983-46ca-a0e3-1021c9f14638" containerName="registry-server" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.145700 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="390604e6-b428-4bd4-b5a7-4e244339148f" containerName="registry-server" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.146704 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.149114 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.158412 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d"] Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.254785 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl7wt\" (UniqueName: \"kubernetes.io/projected/8e5953bd-8302-4918-a2da-911d68e7fc79-kube-api-access-fl7wt\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.254845 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.254883 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.356019 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl7wt\" (UniqueName: \"kubernetes.io/projected/8e5953bd-8302-4918-a2da-911d68e7fc79-kube-api-access-fl7wt\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.356086 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.356129 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.356731 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.356778 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.381333 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl7wt\" (UniqueName: \"kubernetes.io/projected/8e5953bd-8302-4918-a2da-911d68e7fc79-kube-api-access-fl7wt\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.466291 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:30 crc kubenswrapper[4858]: I0320 09:14:30.919142 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d"] Mar 20 09:14:30 crc kubenswrapper[4858]: W0320 09:14:30.949513 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e5953bd_8302_4918_a2da_911d68e7fc79.slice/crio-a7cae71d42cf3fe6b82bba6093db9fe9c41e76bb4ac16fca8841f773609753b6 WatchSource:0}: Error finding container a7cae71d42cf3fe6b82bba6093db9fe9c41e76bb4ac16fca8841f773609753b6: Status 404 returned error can't find the container with id a7cae71d42cf3fe6b82bba6093db9fe9c41e76bb4ac16fca8841f773609753b6 Mar 20 09:14:31 crc kubenswrapper[4858]: I0320 09:14:31.356904 4858 generic.go:334] "Generic (PLEG): container finished" podID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerID="3a0397f826e2ed00cac8859cb9879275b9da36462e4a33e402d8374bddc73c51" exitCode=0 Mar 20 09:14:31 crc kubenswrapper[4858]: I0320 09:14:31.356958 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" event={"ID":"8e5953bd-8302-4918-a2da-911d68e7fc79","Type":"ContainerDied","Data":"3a0397f826e2ed00cac8859cb9879275b9da36462e4a33e402d8374bddc73c51"} Mar 20 09:14:31 crc kubenswrapper[4858]: I0320 09:14:31.357015 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" event={"ID":"8e5953bd-8302-4918-a2da-911d68e7fc79","Type":"ContainerStarted","Data":"a7cae71d42cf3fe6b82bba6093db9fe9c41e76bb4ac16fca8841f773609753b6"} Mar 20 09:14:35 crc kubenswrapper[4858]: I0320 09:14:35.389951 4858 generic.go:334] "Generic (PLEG): container finished" podID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerID="93c673fc30bfadef6c2bfcca0973b50fd63c1cc391c2681e247895535140de41" exitCode=0 Mar 20 09:14:35 crc kubenswrapper[4858]: I0320 09:14:35.390012 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" event={"ID":"8e5953bd-8302-4918-a2da-911d68e7fc79","Type":"ContainerDied","Data":"93c673fc30bfadef6c2bfcca0973b50fd63c1cc391c2681e247895535140de41"} Mar 20 09:14:36 crc kubenswrapper[4858]: I0320 09:14:36.401168 4858 generic.go:334] "Generic (PLEG): container finished" podID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerID="ba8b5ecd956692b863eb80176fd63cfbe7cecddf210409e66073b63d566cfb80" exitCode=0 Mar 20 09:14:36 crc kubenswrapper[4858]: I0320 09:14:36.401307 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" event={"ID":"8e5953bd-8302-4918-a2da-911d68e7fc79","Type":"ContainerDied","Data":"ba8b5ecd956692b863eb80176fd63cfbe7cecddf210409e66073b63d566cfb80"} Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.099192 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-lwq5q" Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.685276 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.875062 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-bundle\") pod \"8e5953bd-8302-4918-a2da-911d68e7fc79\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.875123 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-util\") pod \"8e5953bd-8302-4918-a2da-911d68e7fc79\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.875247 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl7wt\" (UniqueName: \"kubernetes.io/projected/8e5953bd-8302-4918-a2da-911d68e7fc79-kube-api-access-fl7wt\") pod \"8e5953bd-8302-4918-a2da-911d68e7fc79\" (UID: \"8e5953bd-8302-4918-a2da-911d68e7fc79\") " Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.876307 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-bundle" (OuterVolumeSpecName: "bundle") pod "8e5953bd-8302-4918-a2da-911d68e7fc79" (UID: "8e5953bd-8302-4918-a2da-911d68e7fc79"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.884542 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e5953bd-8302-4918-a2da-911d68e7fc79-kube-api-access-fl7wt" (OuterVolumeSpecName: "kube-api-access-fl7wt") pod "8e5953bd-8302-4918-a2da-911d68e7fc79" (UID: "8e5953bd-8302-4918-a2da-911d68e7fc79"). InnerVolumeSpecName "kube-api-access-fl7wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.887180 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-util" (OuterVolumeSpecName: "util") pod "8e5953bd-8302-4918-a2da-911d68e7fc79" (UID: "8e5953bd-8302-4918-a2da-911d68e7fc79"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.976356 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl7wt\" (UniqueName: \"kubernetes.io/projected/8e5953bd-8302-4918-a2da-911d68e7fc79-kube-api-access-fl7wt\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.976388 4858 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:37 crc kubenswrapper[4858]: I0320 09:14:37.976398 4858 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8e5953bd-8302-4918-a2da-911d68e7fc79-util\") on node \"crc\" DevicePath \"\"" Mar 20 09:14:38 crc kubenswrapper[4858]: I0320 09:14:38.420889 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" event={"ID":"8e5953bd-8302-4918-a2da-911d68e7fc79","Type":"ContainerDied","Data":"a7cae71d42cf3fe6b82bba6093db9fe9c41e76bb4ac16fca8841f773609753b6"} Mar 20 09:14:38 crc kubenswrapper[4858]: I0320 09:14:38.420955 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d" Mar 20 09:14:38 crc kubenswrapper[4858]: I0320 09:14:38.420960 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7cae71d42cf3fe6b82bba6093db9fe9c41e76bb4ac16fca8841f773609753b6" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.304844 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v"] Mar 20 09:14:43 crc kubenswrapper[4858]: E0320 09:14:43.306174 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerName="pull" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.306195 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerName="pull" Mar 20 09:14:43 crc kubenswrapper[4858]: E0320 09:14:43.306213 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerName="extract" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.306221 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerName="extract" Mar 20 09:14:43 crc kubenswrapper[4858]: E0320 09:14:43.306237 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerName="util" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.306245 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerName="util" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.306413 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e5953bd-8302-4918-a2da-911d68e7fc79" containerName="extract" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.307071 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.310262 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.311526 4858 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-m8r9d" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.317691 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.332791 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v"] Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.461153 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5wks\" (UniqueName: \"kubernetes.io/projected/d44d4204-44fd-4465-a37b-628325ad79ae-kube-api-access-g5wks\") pod \"cert-manager-operator-controller-manager-66c8bdd694-c992v\" (UID: \"d44d4204-44fd-4465-a37b-628325ad79ae\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.461227 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d44d4204-44fd-4465-a37b-628325ad79ae-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-c992v\" (UID: \"d44d4204-44fd-4465-a37b-628325ad79ae\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.562645 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5wks\" (UniqueName: \"kubernetes.io/projected/d44d4204-44fd-4465-a37b-628325ad79ae-kube-api-access-g5wks\") pod \"cert-manager-operator-controller-manager-66c8bdd694-c992v\" (UID: \"d44d4204-44fd-4465-a37b-628325ad79ae\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.562716 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d44d4204-44fd-4465-a37b-628325ad79ae-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-c992v\" (UID: \"d44d4204-44fd-4465-a37b-628325ad79ae\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.563363 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d44d4204-44fd-4465-a37b-628325ad79ae-tmp\") pod \"cert-manager-operator-controller-manager-66c8bdd694-c992v\" (UID: \"d44d4204-44fd-4465-a37b-628325ad79ae\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.593633 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5wks\" (UniqueName: \"kubernetes.io/projected/d44d4204-44fd-4465-a37b-628325ad79ae-kube-api-access-g5wks\") pod \"cert-manager-operator-controller-manager-66c8bdd694-c992v\" (UID: \"d44d4204-44fd-4465-a37b-628325ad79ae\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.627430 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" Mar 20 09:14:43 crc kubenswrapper[4858]: I0320 09:14:43.931028 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v"] Mar 20 09:14:44 crc kubenswrapper[4858]: I0320 09:14:44.457177 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" event={"ID":"d44d4204-44fd-4465-a37b-628325ad79ae","Type":"ContainerStarted","Data":"308aefd3392c60c45813acd0b65531d1ec36134e411b64f40d7ab224d20e4cdf"} Mar 20 09:14:47 crc kubenswrapper[4858]: I0320 09:14:47.478735 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" event={"ID":"d44d4204-44fd-4465-a37b-628325ad79ae","Type":"ContainerStarted","Data":"ae54117d4a96304b16552d1f946b812c065c3f5d16322eef429d11bf7573a1e5"} Mar 20 09:14:51 crc kubenswrapper[4858]: I0320 09:14:51.951707 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-66c8bdd694-c992v" podStartSLOduration=6.036713625 podStartE2EDuration="8.951686806s" podCreationTimestamp="2026-03-20 09:14:43 +0000 UTC" firstStartedPulling="2026-03-20 09:14:43.9723556 +0000 UTC m=+1065.292773797" lastFinishedPulling="2026-03-20 09:14:46.887328781 +0000 UTC m=+1068.207746978" observedRunningTime="2026-03-20 09:14:47.507740627 +0000 UTC m=+1068.828158844" watchObservedRunningTime="2026-03-20 09:14:51.951686806 +0000 UTC m=+1073.272105013" Mar 20 09:14:51 crc kubenswrapper[4858]: I0320 09:14:51.958291 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-dgqzw"] Mar 20 09:14:51 crc kubenswrapper[4858]: I0320 09:14:51.959280 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:51 crc kubenswrapper[4858]: I0320 09:14:51.963048 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 20 09:14:51 crc kubenswrapper[4858]: I0320 09:14:51.963052 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 20 09:14:51 crc kubenswrapper[4858]: I0320 09:14:51.964677 4858 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-vwx4s" Mar 20 09:14:51 crc kubenswrapper[4858]: I0320 09:14:51.966760 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-dgqzw"] Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.095572 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-dgqzw\" (UID: \"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3\") " pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.095656 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6d56\" (UniqueName: \"kubernetes.io/projected/76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3-kube-api-access-f6d56\") pod \"cert-manager-webhook-6888856db4-dgqzw\" (UID: \"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3\") " pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.196495 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6d56\" (UniqueName: \"kubernetes.io/projected/76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3-kube-api-access-f6d56\") pod \"cert-manager-webhook-6888856db4-dgqzw\" (UID: \"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3\") " pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.196620 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-dgqzw\" (UID: \"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3\") " pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.221705 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6d56\" (UniqueName: \"kubernetes.io/projected/76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3-kube-api-access-f6d56\") pod \"cert-manager-webhook-6888856db4-dgqzw\" (UID: \"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3\") " pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.247204 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-dgqzw\" (UID: \"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3\") " pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.274631 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:52 crc kubenswrapper[4858]: I0320 09:14:52.734633 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-dgqzw"] Mar 20 09:14:53 crc kubenswrapper[4858]: I0320 09:14:53.517063 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" event={"ID":"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3","Type":"ContainerStarted","Data":"58d0a67e7bd4dfd38c3e2b225aeae0c64b11f3ef0ce9da7d37351fd6d506f23e"} Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.459028 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-zxsx6"] Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.460144 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.462485 4858 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-pbc4h" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.485583 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-zxsx6"] Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.551555 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" event={"ID":"76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3","Type":"ContainerStarted","Data":"228ae7dedb739cfc90b156b942b3964753dd37d235dcb538e8654970efa316a6"} Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.551725 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.568512 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" podStartSLOduration=1.968193211 podStartE2EDuration="6.568461874s" podCreationTimestamp="2026-03-20 09:14:51 +0000 UTC" firstStartedPulling="2026-03-20 09:14:52.746746278 +0000 UTC m=+1074.067164485" lastFinishedPulling="2026-03-20 09:14:57.347014951 +0000 UTC m=+1078.667433148" observedRunningTime="2026-03-20 09:14:57.567875332 +0000 UTC m=+1078.888293549" watchObservedRunningTime="2026-03-20 09:14:57.568461874 +0000 UTC m=+1078.888880081" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.580342 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwffv\" (UniqueName: \"kubernetes.io/projected/e7decc95-2f64-45ce-a3d5-679589f5b77a-kube-api-access-zwffv\") pod \"cert-manager-cainjector-5545bd876-zxsx6\" (UID: \"e7decc95-2f64-45ce-a3d5-679589f5b77a\") " pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.580598 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7decc95-2f64-45ce-a3d5-679589f5b77a-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-zxsx6\" (UID: \"e7decc95-2f64-45ce-a3d5-679589f5b77a\") " pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.682518 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7decc95-2f64-45ce-a3d5-679589f5b77a-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-zxsx6\" (UID: \"e7decc95-2f64-45ce-a3d5-679589f5b77a\") " pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.682611 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwffv\" (UniqueName: \"kubernetes.io/projected/e7decc95-2f64-45ce-a3d5-679589f5b77a-kube-api-access-zwffv\") pod \"cert-manager-cainjector-5545bd876-zxsx6\" (UID: \"e7decc95-2f64-45ce-a3d5-679589f5b77a\") " pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.710824 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e7decc95-2f64-45ce-a3d5-679589f5b77a-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-zxsx6\" (UID: \"e7decc95-2f64-45ce-a3d5-679589f5b77a\") " pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.718234 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwffv\" (UniqueName: \"kubernetes.io/projected/e7decc95-2f64-45ce-a3d5-679589f5b77a-kube-api-access-zwffv\") pod \"cert-manager-cainjector-5545bd876-zxsx6\" (UID: \"e7decc95-2f64-45ce-a3d5-679589f5b77a\") " pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:57 crc kubenswrapper[4858]: I0320 09:14:57.807749 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" Mar 20 09:14:58 crc kubenswrapper[4858]: I0320 09:14:58.300556 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-zxsx6"] Mar 20 09:14:58 crc kubenswrapper[4858]: W0320 09:14:58.306178 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7decc95_2f64_45ce_a3d5_679589f5b77a.slice/crio-e831306df2ef4d70c68f3ad3ee6a603afd9b187e45a1e41d5c8a33b3c2ab15ee WatchSource:0}: Error finding container e831306df2ef4d70c68f3ad3ee6a603afd9b187e45a1e41d5c8a33b3c2ab15ee: Status 404 returned error can't find the container with id e831306df2ef4d70c68f3ad3ee6a603afd9b187e45a1e41d5c8a33b3c2ab15ee Mar 20 09:14:58 crc kubenswrapper[4858]: I0320 09:14:58.562685 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" event={"ID":"e7decc95-2f64-45ce-a3d5-679589f5b77a","Type":"ContainerStarted","Data":"2f1bba9f313aebce3c532b7524a11942d7c598d5859c4e2a761c9d832bae7be5"} Mar 20 09:14:58 crc kubenswrapper[4858]: I0320 09:14:58.563194 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" event={"ID":"e7decc95-2f64-45ce-a3d5-679589f5b77a","Type":"ContainerStarted","Data":"e831306df2ef4d70c68f3ad3ee6a603afd9b187e45a1e41d5c8a33b3c2ab15ee"} Mar 20 09:14:58 crc kubenswrapper[4858]: I0320 09:14:58.587189 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-zxsx6" podStartSLOduration=1.587162847 podStartE2EDuration="1.587162847s" podCreationTimestamp="2026-03-20 09:14:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:14:58.579039232 +0000 UTC m=+1079.899457439" watchObservedRunningTime="2026-03-20 09:14:58.587162847 +0000 UTC m=+1079.907581044" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.132108 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh"] Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.133524 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.137135 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.137812 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.141949 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh"] Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.322540 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-config-volume\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.322652 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-secret-volume\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.322689 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnss7\" (UniqueName: \"kubernetes.io/projected/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-kube-api-access-xnss7\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.424688 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-secret-volume\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.424746 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnss7\" (UniqueName: \"kubernetes.io/projected/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-kube-api-access-xnss7\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.424813 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-config-volume\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.425994 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-config-volume\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.443198 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-secret-volume\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.443344 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnss7\" (UniqueName: \"kubernetes.io/projected/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-kube-api-access-xnss7\") pod \"collect-profiles-29566635-pwrbh\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.500065 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:00 crc kubenswrapper[4858]: I0320 09:15:00.739996 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh"] Mar 20 09:15:01 crc kubenswrapper[4858]: I0320 09:15:01.583417 4858 generic.go:334] "Generic (PLEG): container finished" podID="8ea3444d-144e-45fb-aaff-b4fb8ee5d758" containerID="ff70c71525e7dc7d3f3ba3c840b020a16e0c68d2c1a0af82344e85ec7406b27a" exitCode=0 Mar 20 09:15:01 crc kubenswrapper[4858]: I0320 09:15:01.583879 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" event={"ID":"8ea3444d-144e-45fb-aaff-b4fb8ee5d758","Type":"ContainerDied","Data":"ff70c71525e7dc7d3f3ba3c840b020a16e0c68d2c1a0af82344e85ec7406b27a"} Mar 20 09:15:01 crc kubenswrapper[4858]: I0320 09:15:01.583923 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" event={"ID":"8ea3444d-144e-45fb-aaff-b4fb8ee5d758","Type":"ContainerStarted","Data":"7e6bcc423d076e706fe111614899aefad50c7dba237c1a9ad91a7285decbf97d"} Mar 20 09:15:02 crc kubenswrapper[4858]: I0320 09:15:02.279144 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-dgqzw" Mar 20 09:15:02 crc kubenswrapper[4858]: I0320 09:15:02.891642 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.072529 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-secret-volume\") pod \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.072678 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-config-volume\") pod \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.073051 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnss7\" (UniqueName: \"kubernetes.io/projected/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-kube-api-access-xnss7\") pod \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\" (UID: \"8ea3444d-144e-45fb-aaff-b4fb8ee5d758\") " Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.073927 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-config-volume" (OuterVolumeSpecName: "config-volume") pod "8ea3444d-144e-45fb-aaff-b4fb8ee5d758" (UID: "8ea3444d-144e-45fb-aaff-b4fb8ee5d758"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.080150 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-kube-api-access-xnss7" (OuterVolumeSpecName: "kube-api-access-xnss7") pod "8ea3444d-144e-45fb-aaff-b4fb8ee5d758" (UID: "8ea3444d-144e-45fb-aaff-b4fb8ee5d758"). InnerVolumeSpecName "kube-api-access-xnss7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.082668 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8ea3444d-144e-45fb-aaff-b4fb8ee5d758" (UID: "8ea3444d-144e-45fb-aaff-b4fb8ee5d758"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.175763 4858 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.175825 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnss7\" (UniqueName: \"kubernetes.io/projected/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-kube-api-access-xnss7\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.175838 4858 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8ea3444d-144e-45fb-aaff-b4fb8ee5d758-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.600252 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" event={"ID":"8ea3444d-144e-45fb-aaff-b4fb8ee5d758","Type":"ContainerDied","Data":"7e6bcc423d076e706fe111614899aefad50c7dba237c1a9ad91a7285decbf97d"} Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.600305 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566635-pwrbh" Mar 20 09:15:03 crc kubenswrapper[4858]: I0320 09:15:03.600362 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e6bcc423d076e706fe111614899aefad50c7dba237c1a9ad91a7285decbf97d" Mar 20 09:15:07 crc kubenswrapper[4858]: I0320 09:15:07.891981 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:15:07 crc kubenswrapper[4858]: I0320 09:15:07.892488 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.824662 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-s5c5l"] Mar 20 09:15:10 crc kubenswrapper[4858]: E0320 09:15:10.825085 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ea3444d-144e-45fb-aaff-b4fb8ee5d758" containerName="collect-profiles" Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.825101 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ea3444d-144e-45fb-aaff-b4fb8ee5d758" containerName="collect-profiles" Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.825299 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ea3444d-144e-45fb-aaff-b4fb8ee5d758" containerName="collect-profiles" Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.826023 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.828462 4858 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-bqskv" Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.833010 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-s5c5l"] Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.908018 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/714bde66-3be8-485f-af48-b32f2177ae05-bound-sa-token\") pod \"cert-manager-545d4d4674-s5c5l\" (UID: \"714bde66-3be8-485f-af48-b32f2177ae05\") " pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:10 crc kubenswrapper[4858]: I0320 09:15:10.908115 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9f4g\" (UniqueName: \"kubernetes.io/projected/714bde66-3be8-485f-af48-b32f2177ae05-kube-api-access-f9f4g\") pod \"cert-manager-545d4d4674-s5c5l\" (UID: \"714bde66-3be8-485f-af48-b32f2177ae05\") " pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.009233 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/714bde66-3be8-485f-af48-b32f2177ae05-bound-sa-token\") pod \"cert-manager-545d4d4674-s5c5l\" (UID: \"714bde66-3be8-485f-af48-b32f2177ae05\") " pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.009804 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9f4g\" (UniqueName: \"kubernetes.io/projected/714bde66-3be8-485f-af48-b32f2177ae05-kube-api-access-f9f4g\") pod \"cert-manager-545d4d4674-s5c5l\" (UID: \"714bde66-3be8-485f-af48-b32f2177ae05\") " pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.040669 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9f4g\" (UniqueName: \"kubernetes.io/projected/714bde66-3be8-485f-af48-b32f2177ae05-kube-api-access-f9f4g\") pod \"cert-manager-545d4d4674-s5c5l\" (UID: \"714bde66-3be8-485f-af48-b32f2177ae05\") " pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.041286 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/714bde66-3be8-485f-af48-b32f2177ae05-bound-sa-token\") pod \"cert-manager-545d4d4674-s5c5l\" (UID: \"714bde66-3be8-485f-af48-b32f2177ae05\") " pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.153064 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-s5c5l" Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.394874 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-s5c5l"] Mar 20 09:15:11 crc kubenswrapper[4858]: W0320 09:15:11.401622 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod714bde66_3be8_485f_af48_b32f2177ae05.slice/crio-9d807832ebb2eb3181f05697f099b1fffc4ee3ebc49a69c0f525665bc7b05739 WatchSource:0}: Error finding container 9d807832ebb2eb3181f05697f099b1fffc4ee3ebc49a69c0f525665bc7b05739: Status 404 returned error can't find the container with id 9d807832ebb2eb3181f05697f099b1fffc4ee3ebc49a69c0f525665bc7b05739 Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.658552 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-s5c5l" event={"ID":"714bde66-3be8-485f-af48-b32f2177ae05","Type":"ContainerStarted","Data":"351c1bd2249f388d097d8def90731b5dfdf2324286fa9ce66ceebbaf372d2148"} Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.658834 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-s5c5l" event={"ID":"714bde66-3be8-485f-af48-b32f2177ae05","Type":"ContainerStarted","Data":"9d807832ebb2eb3181f05697f099b1fffc4ee3ebc49a69c0f525665bc7b05739"} Mar 20 09:15:11 crc kubenswrapper[4858]: I0320 09:15:11.679545 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-s5c5l" podStartSLOduration=1.679527054 podStartE2EDuration="1.679527054s" podCreationTimestamp="2026-03-20 09:15:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:15:11.675207892 +0000 UTC m=+1092.995626099" watchObservedRunningTime="2026-03-20 09:15:11.679527054 +0000 UTC m=+1092.999945251" Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.714126 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-s44rj"] Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.715703 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.718094 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-h776f" Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.721294 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.721775 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.722325 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s44rj"] Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.806326 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmcfp\" (UniqueName: \"kubernetes.io/projected/688b14a3-bd9d-45d9-8adc-c65703cfcd9d-kube-api-access-jmcfp\") pod \"openstack-operator-index-s44rj\" (UID: \"688b14a3-bd9d-45d9-8adc-c65703cfcd9d\") " pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.908142 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmcfp\" (UniqueName: \"kubernetes.io/projected/688b14a3-bd9d-45d9-8adc-c65703cfcd9d-kube-api-access-jmcfp\") pod \"openstack-operator-index-s44rj\" (UID: \"688b14a3-bd9d-45d9-8adc-c65703cfcd9d\") " pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:15 crc kubenswrapper[4858]: I0320 09:15:15.929352 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmcfp\" (UniqueName: \"kubernetes.io/projected/688b14a3-bd9d-45d9-8adc-c65703cfcd9d-kube-api-access-jmcfp\") pod \"openstack-operator-index-s44rj\" (UID: \"688b14a3-bd9d-45d9-8adc-c65703cfcd9d\") " pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:16 crc kubenswrapper[4858]: I0320 09:15:16.075190 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:16 crc kubenswrapper[4858]: I0320 09:15:16.300899 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-s44rj"] Mar 20 09:15:16 crc kubenswrapper[4858]: W0320 09:15:16.313859 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod688b14a3_bd9d_45d9_8adc_c65703cfcd9d.slice/crio-1e4e300f30e0e6009b0abfd6d43ef31b3eeff2c76c72ba42e930ccaf1e6670f3 WatchSource:0}: Error finding container 1e4e300f30e0e6009b0abfd6d43ef31b3eeff2c76c72ba42e930ccaf1e6670f3: Status 404 returned error can't find the container with id 1e4e300f30e0e6009b0abfd6d43ef31b3eeff2c76c72ba42e930ccaf1e6670f3 Mar 20 09:15:16 crc kubenswrapper[4858]: I0320 09:15:16.694081 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s44rj" event={"ID":"688b14a3-bd9d-45d9-8adc-c65703cfcd9d","Type":"ContainerStarted","Data":"1e4e300f30e0e6009b0abfd6d43ef31b3eeff2c76c72ba42e930ccaf1e6670f3"} Mar 20 09:15:19 crc kubenswrapper[4858]: I0320 09:15:19.718533 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-s44rj" event={"ID":"688b14a3-bd9d-45d9-8adc-c65703cfcd9d","Type":"ContainerStarted","Data":"95b5c66f5f1ed50957515e71e4e7df4c018ed54cd3dbbc3180429355326e4a3e"} Mar 20 09:15:19 crc kubenswrapper[4858]: I0320 09:15:19.739131 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-s44rj" podStartSLOduration=2.46171189 podStartE2EDuration="4.739104435s" podCreationTimestamp="2026-03-20 09:15:15 +0000 UTC" firstStartedPulling="2026-03-20 09:15:16.31552777 +0000 UTC m=+1097.635945967" lastFinishedPulling="2026-03-20 09:15:18.592920315 +0000 UTC m=+1099.913338512" observedRunningTime="2026-03-20 09:15:19.737105463 +0000 UTC m=+1101.057523670" watchObservedRunningTime="2026-03-20 09:15:19.739104435 +0000 UTC m=+1101.059522642" Mar 20 09:15:26 crc kubenswrapper[4858]: I0320 09:15:26.082396 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:26 crc kubenswrapper[4858]: I0320 09:15:26.083381 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:26 crc kubenswrapper[4858]: I0320 09:15:26.120332 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:26 crc kubenswrapper[4858]: I0320 09:15:26.807555 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-s44rj" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.713779 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph"] Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.715626 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.717869 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-hgsq9" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.726216 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph"] Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.763888 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j57fx\" (UniqueName: \"kubernetes.io/projected/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-kube-api-access-j57fx\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.763991 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-util\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.764044 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-bundle\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.866446 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-util\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.866866 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-bundle\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.866980 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j57fx\" (UniqueName: \"kubernetes.io/projected/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-kube-api-access-j57fx\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.867172 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-util\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.867467 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-bundle\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:31 crc kubenswrapper[4858]: I0320 09:15:31.891864 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j57fx\" (UniqueName: \"kubernetes.io/projected/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-kube-api-access-j57fx\") pod \"b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:32 crc kubenswrapper[4858]: I0320 09:15:32.036143 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:32 crc kubenswrapper[4858]: I0320 09:15:32.329005 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph"] Mar 20 09:15:32 crc kubenswrapper[4858]: I0320 09:15:32.828196 4858 generic.go:334] "Generic (PLEG): container finished" podID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerID="0387251f9526ed7df3150b986bb2e7511ba3f0c7927e1ca3256ed767bd453495" exitCode=0 Mar 20 09:15:32 crc kubenswrapper[4858]: I0320 09:15:32.828409 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" event={"ID":"6c4644e0-05b7-4776-b0ae-d45502e6f6b4","Type":"ContainerDied","Data":"0387251f9526ed7df3150b986bb2e7511ba3f0c7927e1ca3256ed767bd453495"} Mar 20 09:15:32 crc kubenswrapper[4858]: I0320 09:15:32.828525 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" event={"ID":"6c4644e0-05b7-4776-b0ae-d45502e6f6b4","Type":"ContainerStarted","Data":"8f71f8193f2cb5f848744cd791357e1d1292399c1938888733044964f0904190"} Mar 20 09:15:33 crc kubenswrapper[4858]: I0320 09:15:33.837728 4858 generic.go:334] "Generic (PLEG): container finished" podID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerID="1de3fc02904b1f53002ab4bc6833e341cd59ac5eadffe7c0a80aac2f2c27e96d" exitCode=0 Mar 20 09:15:33 crc kubenswrapper[4858]: I0320 09:15:33.837854 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" event={"ID":"6c4644e0-05b7-4776-b0ae-d45502e6f6b4","Type":"ContainerDied","Data":"1de3fc02904b1f53002ab4bc6833e341cd59ac5eadffe7c0a80aac2f2c27e96d"} Mar 20 09:15:34 crc kubenswrapper[4858]: E0320 09:15:34.150496 4858 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c4644e0_05b7_4776_b0ae_d45502e6f6b4.slice/crio-db7823ba8f7bd99600a93916481737dbf2c9e498fea7d242ec93bf931f49f75e.scope\": RecentStats: unable to find data in memory cache]" Mar 20 09:15:34 crc kubenswrapper[4858]: I0320 09:15:34.852659 4858 generic.go:334] "Generic (PLEG): container finished" podID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerID="db7823ba8f7bd99600a93916481737dbf2c9e498fea7d242ec93bf931f49f75e" exitCode=0 Mar 20 09:15:34 crc kubenswrapper[4858]: I0320 09:15:34.852975 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" event={"ID":"6c4644e0-05b7-4776-b0ae-d45502e6f6b4","Type":"ContainerDied","Data":"db7823ba8f7bd99600a93916481737dbf2c9e498fea7d242ec93bf931f49f75e"} Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.118179 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.242785 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j57fx\" (UniqueName: \"kubernetes.io/projected/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-kube-api-access-j57fx\") pod \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.242849 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-util\") pod \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.242952 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-bundle\") pod \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\" (UID: \"6c4644e0-05b7-4776-b0ae-d45502e6f6b4\") " Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.244176 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-bundle" (OuterVolumeSpecName: "bundle") pod "6c4644e0-05b7-4776-b0ae-d45502e6f6b4" (UID: "6c4644e0-05b7-4776-b0ae-d45502e6f6b4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.252238 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-kube-api-access-j57fx" (OuterVolumeSpecName: "kube-api-access-j57fx") pod "6c4644e0-05b7-4776-b0ae-d45502e6f6b4" (UID: "6c4644e0-05b7-4776-b0ae-d45502e6f6b4"). InnerVolumeSpecName "kube-api-access-j57fx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.276936 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-util" (OuterVolumeSpecName: "util") pod "6c4644e0-05b7-4776-b0ae-d45502e6f6b4" (UID: "6c4644e0-05b7-4776-b0ae-d45502e6f6b4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.345276 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j57fx\" (UniqueName: \"kubernetes.io/projected/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-kube-api-access-j57fx\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.345651 4858 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-util\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.345781 4858 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c4644e0-05b7-4776-b0ae-d45502e6f6b4-bundle\") on node \"crc\" DevicePath \"\"" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.873859 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" event={"ID":"6c4644e0-05b7-4776-b0ae-d45502e6f6b4","Type":"ContainerDied","Data":"8f71f8193f2cb5f848744cd791357e1d1292399c1938888733044964f0904190"} Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.873937 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f71f8193f2cb5f848744cd791357e1d1292399c1938888733044964f0904190" Mar 20 09:15:36 crc kubenswrapper[4858]: I0320 09:15:36.874065 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph" Mar 20 09:15:37 crc kubenswrapper[4858]: I0320 09:15:37.890694 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:15:37 crc kubenswrapper[4858]: I0320 09:15:37.891159 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.325064 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln"] Mar 20 09:15:39 crc kubenswrapper[4858]: E0320 09:15:39.325403 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerName="util" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.325421 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerName="util" Mar 20 09:15:39 crc kubenswrapper[4858]: E0320 09:15:39.325450 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerName="pull" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.325458 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerName="pull" Mar 20 09:15:39 crc kubenswrapper[4858]: E0320 09:15:39.325467 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerName="extract" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.325475 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerName="extract" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.325616 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c4644e0-05b7-4776-b0ae-d45502e6f6b4" containerName="extract" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.326140 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.328571 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-wwzkj" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.363639 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln"] Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.415024 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctp2j\" (UniqueName: \"kubernetes.io/projected/23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de-kube-api-access-ctp2j\") pod \"openstack-operator-controller-init-9df8dd5fd-7ssln\" (UID: \"23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de\") " pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.517289 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctp2j\" (UniqueName: \"kubernetes.io/projected/23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de-kube-api-access-ctp2j\") pod \"openstack-operator-controller-init-9df8dd5fd-7ssln\" (UID: \"23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de\") " pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.540521 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctp2j\" (UniqueName: \"kubernetes.io/projected/23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de-kube-api-access-ctp2j\") pod \"openstack-operator-controller-init-9df8dd5fd-7ssln\" (UID: \"23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de\") " pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" Mar 20 09:15:39 crc kubenswrapper[4858]: I0320 09:15:39.644227 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" Mar 20 09:15:40 crc kubenswrapper[4858]: I0320 09:15:40.083233 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln"] Mar 20 09:15:40 crc kubenswrapper[4858]: I0320 09:15:40.910583 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" event={"ID":"23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de","Type":"ContainerStarted","Data":"ac814df904dcbc216539529001de87ff8cd0f986a789f14322b9ddbac7a658d5"} Mar 20 09:15:44 crc kubenswrapper[4858]: I0320 09:15:44.949985 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" event={"ID":"23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de","Type":"ContainerStarted","Data":"9464132fb96ab1045d7ab96db35e2e5c35fe9474d0a60c55e28a9c8000517f63"} Mar 20 09:15:44 crc kubenswrapper[4858]: I0320 09:15:44.950759 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" Mar 20 09:15:44 crc kubenswrapper[4858]: I0320 09:15:44.980890 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" podStartSLOduration=1.597814392 podStartE2EDuration="5.980861385s" podCreationTimestamp="2026-03-20 09:15:39 +0000 UTC" firstStartedPulling="2026-03-20 09:15:40.114994152 +0000 UTC m=+1121.435412349" lastFinishedPulling="2026-03-20 09:15:44.498041145 +0000 UTC m=+1125.818459342" observedRunningTime="2026-03-20 09:15:44.974521277 +0000 UTC m=+1126.294939484" watchObservedRunningTime="2026-03-20 09:15:44.980861385 +0000 UTC m=+1126.301279582" Mar 20 09:15:49 crc kubenswrapper[4858]: I0320 09:15:49.649209 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-9df8dd5fd-7ssln" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.136221 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566636-8b64v"] Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.138537 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-8b64v" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.140182 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.141234 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.141397 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.147713 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-8b64v"] Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.239784 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpvsp\" (UniqueName: \"kubernetes.io/projected/d1276f93-9455-48ff-82c1-3ddba162054f-kube-api-access-mpvsp\") pod \"auto-csr-approver-29566636-8b64v\" (UID: \"d1276f93-9455-48ff-82c1-3ddba162054f\") " pod="openshift-infra/auto-csr-approver-29566636-8b64v" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.341001 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpvsp\" (UniqueName: \"kubernetes.io/projected/d1276f93-9455-48ff-82c1-3ddba162054f-kube-api-access-mpvsp\") pod \"auto-csr-approver-29566636-8b64v\" (UID: \"d1276f93-9455-48ff-82c1-3ddba162054f\") " pod="openshift-infra/auto-csr-approver-29566636-8b64v" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.364566 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpvsp\" (UniqueName: \"kubernetes.io/projected/d1276f93-9455-48ff-82c1-3ddba162054f-kube-api-access-mpvsp\") pod \"auto-csr-approver-29566636-8b64v\" (UID: \"d1276f93-9455-48ff-82c1-3ddba162054f\") " pod="openshift-infra/auto-csr-approver-29566636-8b64v" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.465338 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-8b64v" Mar 20 09:16:00 crc kubenswrapper[4858]: I0320 09:16:00.694360 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-8b64v"] Mar 20 09:16:00 crc kubenswrapper[4858]: W0320 09:16:00.704879 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1276f93_9455_48ff_82c1_3ddba162054f.slice/crio-b720b0b268a3418f51bad98ae5aef801899402e78f2d9e4d5ab2cc58ba0ebc8f WatchSource:0}: Error finding container b720b0b268a3418f51bad98ae5aef801899402e78f2d9e4d5ab2cc58ba0ebc8f: Status 404 returned error can't find the container with id b720b0b268a3418f51bad98ae5aef801899402e78f2d9e4d5ab2cc58ba0ebc8f Mar 20 09:16:01 crc kubenswrapper[4858]: I0320 09:16:01.060613 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566636-8b64v" event={"ID":"d1276f93-9455-48ff-82c1-3ddba162054f","Type":"ContainerStarted","Data":"b720b0b268a3418f51bad98ae5aef801899402e78f2d9e4d5ab2cc58ba0ebc8f"} Mar 20 09:16:03 crc kubenswrapper[4858]: I0320 09:16:03.077242 4858 generic.go:334] "Generic (PLEG): container finished" podID="d1276f93-9455-48ff-82c1-3ddba162054f" containerID="3de46f8786541609f7420ade58699a2ad681e40cebcc0404f7bd1ca364f3a1d8" exitCode=0 Mar 20 09:16:03 crc kubenswrapper[4858]: I0320 09:16:03.077638 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566636-8b64v" event={"ID":"d1276f93-9455-48ff-82c1-3ddba162054f","Type":"ContainerDied","Data":"3de46f8786541609f7420ade58699a2ad681e40cebcc0404f7bd1ca364f3a1d8"} Mar 20 09:16:04 crc kubenswrapper[4858]: I0320 09:16:04.370024 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-8b64v" Mar 20 09:16:04 crc kubenswrapper[4858]: I0320 09:16:04.415978 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpvsp\" (UniqueName: \"kubernetes.io/projected/d1276f93-9455-48ff-82c1-3ddba162054f-kube-api-access-mpvsp\") pod \"d1276f93-9455-48ff-82c1-3ddba162054f\" (UID: \"d1276f93-9455-48ff-82c1-3ddba162054f\") " Mar 20 09:16:04 crc kubenswrapper[4858]: I0320 09:16:04.421488 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1276f93-9455-48ff-82c1-3ddba162054f-kube-api-access-mpvsp" (OuterVolumeSpecName: "kube-api-access-mpvsp") pod "d1276f93-9455-48ff-82c1-3ddba162054f" (UID: "d1276f93-9455-48ff-82c1-3ddba162054f"). InnerVolumeSpecName "kube-api-access-mpvsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:16:04 crc kubenswrapper[4858]: I0320 09:16:04.517912 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpvsp\" (UniqueName: \"kubernetes.io/projected/d1276f93-9455-48ff-82c1-3ddba162054f-kube-api-access-mpvsp\") on node \"crc\" DevicePath \"\"" Mar 20 09:16:05 crc kubenswrapper[4858]: I0320 09:16:05.091622 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566636-8b64v" event={"ID":"d1276f93-9455-48ff-82c1-3ddba162054f","Type":"ContainerDied","Data":"b720b0b268a3418f51bad98ae5aef801899402e78f2d9e4d5ab2cc58ba0ebc8f"} Mar 20 09:16:05 crc kubenswrapper[4858]: I0320 09:16:05.091672 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b720b0b268a3418f51bad98ae5aef801899402e78f2d9e4d5ab2cc58ba0ebc8f" Mar 20 09:16:05 crc kubenswrapper[4858]: I0320 09:16:05.091728 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566636-8b64v" Mar 20 09:16:05 crc kubenswrapper[4858]: I0320 09:16:05.416374 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-lp6j8"] Mar 20 09:16:05 crc kubenswrapper[4858]: I0320 09:16:05.422090 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566630-lp6j8"] Mar 20 09:16:06 crc kubenswrapper[4858]: I0320 09:16:06.076579 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e4e001-1a9c-4669-b56a-d0d8abfe327d" path="/var/lib/kubelet/pods/63e4e001-1a9c-4669-b56a-d0d8abfe327d/volumes" Mar 20 09:16:07 crc kubenswrapper[4858]: I0320 09:16:07.890490 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:16:07 crc kubenswrapper[4858]: I0320 09:16:07.890939 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:16:07 crc kubenswrapper[4858]: I0320 09:16:07.891008 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:16:07 crc kubenswrapper[4858]: I0320 09:16:07.891950 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bef7d78bb90262eb2557357139a02f6a23b1e0a616279703c46e019d97babf79"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:16:07 crc kubenswrapper[4858]: I0320 09:16:07.892019 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://bef7d78bb90262eb2557357139a02f6a23b1e0a616279703c46e019d97babf79" gracePeriod=600 Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.019406 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv"] Mar 20 09:16:08 crc kubenswrapper[4858]: E0320 09:16:08.020586 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1276f93-9455-48ff-82c1-3ddba162054f" containerName="oc" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.020611 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1276f93-9455-48ff-82c1-3ddba162054f" containerName="oc" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.020762 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1276f93-9455-48ff-82c1-3ddba162054f" containerName="oc" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.021243 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.024252 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-6t69h" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.025841 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.026860 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.028801 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-lv6vj" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.129778 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.136238 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.139504 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="bef7d78bb90262eb2557357139a02f6a23b1e0a616279703c46e019d97babf79" exitCode=0 Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.139578 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"bef7d78bb90262eb2557357139a02f6a23b1e0a616279703c46e019d97babf79"} Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.139623 4858 scope.go:117] "RemoveContainer" containerID="2e450803b001f2a0183f7a90ae4b9b24f8c995b72aa498eab30eafb0ce280f7d" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.140503 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.157371 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.161775 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-r2nfx" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.172728 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.174092 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.180405 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gkmgj" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.187038 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.187122 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwlxj\" (UniqueName: \"kubernetes.io/projected/0e2d5479-7826-4759-99ed-c3775a01035c-kube-api-access-zwlxj\") pod \"barbican-operator-controller-manager-59bc569d95-zk2d4\" (UID: \"0e2d5479-7826-4759-99ed-c3775a01035c\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.187236 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfzsh\" (UniqueName: \"kubernetes.io/projected/cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e-kube-api-access-lfzsh\") pod \"cinder-operator-controller-manager-8d58dc466-7c7lv\" (UID: \"cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.208029 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.218590 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.219737 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.221426 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-z2kp7" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.244544 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.251847 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.255684 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-wl9xv" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.260346 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.283746 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.294156 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v97hw\" (UniqueName: \"kubernetes.io/projected/81033dc9-91dc-4bff-b69d-e1171f82b83c-kube-api-access-v97hw\") pod \"glance-operator-controller-manager-79df6bcc97-bb9mf\" (UID: \"81033dc9-91dc-4bff-b69d-e1171f82b83c\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.294202 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfxlb\" (UniqueName: \"kubernetes.io/projected/feae143f-10cc-4412-9c35-70499e77b5bd-kube-api-access-sfxlb\") pod \"heat-operator-controller-manager-67dd5f86f5-ng56m\" (UID: \"feae143f-10cc-4412-9c35-70499e77b5bd\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.294231 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfzsh\" (UniqueName: \"kubernetes.io/projected/cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e-kube-api-access-lfzsh\") pod \"cinder-operator-controller-manager-8d58dc466-7c7lv\" (UID: \"cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.294271 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86rzd\" (UniqueName: \"kubernetes.io/projected/e2a119a6-367e-4a85-8617-e83d9f6a23b7-kube-api-access-86rzd\") pod \"designate-operator-controller-manager-588d4d986b-8bwmb\" (UID: \"e2a119a6-367e-4a85-8617-e83d9f6a23b7\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.294356 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwlxj\" (UniqueName: \"kubernetes.io/projected/0e2d5479-7826-4759-99ed-c3775a01035c-kube-api-access-zwlxj\") pod \"barbican-operator-controller-manager-59bc569d95-zk2d4\" (UID: \"0e2d5479-7826-4759-99ed-c3775a01035c\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.295785 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.301940 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-sjtxf" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.318855 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.326772 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.329983 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.334485 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfzsh\" (UniqueName: \"kubernetes.io/projected/cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e-kube-api-access-lfzsh\") pod \"cinder-operator-controller-manager-8d58dc466-7c7lv\" (UID: \"cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.337669 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-xds4v" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.337943 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.340465 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwlxj\" (UniqueName: \"kubernetes.io/projected/0e2d5479-7826-4759-99ed-c3775a01035c-kube-api-access-zwlxj\") pod \"barbican-operator-controller-manager-59bc569d95-zk2d4\" (UID: \"0e2d5479-7826-4759-99ed-c3775a01035c\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.357522 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.358483 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.368802 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-hbgmv" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.381994 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.383076 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.385557 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hfdcd" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.391137 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.395092 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrp4w\" (UniqueName: \"kubernetes.io/projected/1e41ad93-0360-4a67-a155-ee05abfbcee1-kube-api-access-jrp4w\") pod \"horizon-operator-controller-manager-8464cc45fb-sk6sh\" (UID: \"1e41ad93-0360-4a67-a155-ee05abfbcee1\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.394669 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.395141 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v97hw\" (UniqueName: \"kubernetes.io/projected/81033dc9-91dc-4bff-b69d-e1171f82b83c-kube-api-access-v97hw\") pod \"glance-operator-controller-manager-79df6bcc97-bb9mf\" (UID: \"81033dc9-91dc-4bff-b69d-e1171f82b83c\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.395177 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfxlb\" (UniqueName: \"kubernetes.io/projected/feae143f-10cc-4412-9c35-70499e77b5bd-kube-api-access-sfxlb\") pod \"heat-operator-controller-manager-67dd5f86f5-ng56m\" (UID: \"feae143f-10cc-4412-9c35-70499e77b5bd\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.395204 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.395227 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86rzd\" (UniqueName: \"kubernetes.io/projected/e2a119a6-367e-4a85-8617-e83d9f6a23b7-kube-api-access-86rzd\") pod \"designate-operator-controller-manager-588d4d986b-8bwmb\" (UID: \"e2a119a6-367e-4a85-8617-e83d9f6a23b7\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.395265 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vk28\" (UniqueName: \"kubernetes.io/projected/a0e45ec4-0059-41b8-8897-8004d4adb9da-kube-api-access-4vk28\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.395301 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkvf7\" (UniqueName: \"kubernetes.io/projected/1adeed8a-d531-46ee-b037-ea137468e026-kube-api-access-xkvf7\") pod \"ironic-operator-controller-manager-6f787dddc9-cb6dz\" (UID: \"1adeed8a-d531-46ee-b037-ea137468e026\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.410943 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.436899 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86rzd\" (UniqueName: \"kubernetes.io/projected/e2a119a6-367e-4a85-8617-e83d9f6a23b7-kube-api-access-86rzd\") pod \"designate-operator-controller-manager-588d4d986b-8bwmb\" (UID: \"e2a119a6-367e-4a85-8617-e83d9f6a23b7\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.441353 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfxlb\" (UniqueName: \"kubernetes.io/projected/feae143f-10cc-4412-9c35-70499e77b5bd-kube-api-access-sfxlb\") pod \"heat-operator-controller-manager-67dd5f86f5-ng56m\" (UID: \"feae143f-10cc-4412-9c35-70499e77b5bd\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.442061 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v97hw\" (UniqueName: \"kubernetes.io/projected/81033dc9-91dc-4bff-b69d-e1171f82b83c-kube-api-access-v97hw\") pod \"glance-operator-controller-manager-79df6bcc97-bb9mf\" (UID: \"81033dc9-91dc-4bff-b69d-e1171f82b83c\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.447649 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.448997 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.462039 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-8sjbp" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.488413 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.491993 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.500540 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrp4w\" (UniqueName: \"kubernetes.io/projected/1e41ad93-0360-4a67-a155-ee05abfbcee1-kube-api-access-jrp4w\") pod \"horizon-operator-controller-manager-8464cc45fb-sk6sh\" (UID: \"1e41ad93-0360-4a67-a155-ee05abfbcee1\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.500600 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.500651 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vk28\" (UniqueName: \"kubernetes.io/projected/a0e45ec4-0059-41b8-8897-8004d4adb9da-kube-api-access-4vk28\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.500696 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldh7q\" (UniqueName: \"kubernetes.io/projected/3dfa5be0-6028-4091-8129-71fa07ab93ac-kube-api-access-ldh7q\") pod \"keystone-operator-controller-manager-768b96df4c-bl6x4\" (UID: \"3dfa5be0-6028-4091-8129-71fa07ab93ac\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.500731 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkvf7\" (UniqueName: \"kubernetes.io/projected/1adeed8a-d531-46ee-b037-ea137468e026-kube-api-access-xkvf7\") pod \"ironic-operator-controller-manager-6f787dddc9-cb6dz\" (UID: \"1adeed8a-d531-46ee-b037-ea137468e026\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.500757 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvsws\" (UniqueName: \"kubernetes.io/projected/71693924-c907-41df-b4ab-cbd9bfb7f97d-kube-api-access-hvsws\") pod \"manila-operator-controller-manager-55f864c847-sv6f4\" (UID: \"71693924-c907-41df-b4ab-cbd9bfb7f97d\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" Mar 20 09:16:08 crc kubenswrapper[4858]: E0320 09:16:08.500893 4858 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:08 crc kubenswrapper[4858]: E0320 09:16:08.500962 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert podName:a0e45ec4-0059-41b8-8897-8004d4adb9da nodeName:}" failed. No retries permitted until 2026-03-20 09:16:09.000941156 +0000 UTC m=+1150.321359343 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert") pod "infra-operator-controller-manager-577ccd856-nw2q9" (UID: "a0e45ec4-0059-41b8-8897-8004d4adb9da") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.531488 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.533877 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrp4w\" (UniqueName: \"kubernetes.io/projected/1e41ad93-0360-4a67-a155-ee05abfbcee1-kube-api-access-jrp4w\") pod \"horizon-operator-controller-manager-8464cc45fb-sk6sh\" (UID: \"1e41ad93-0360-4a67-a155-ee05abfbcee1\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.533929 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.543844 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vk28\" (UniqueName: \"kubernetes.io/projected/a0e45ec4-0059-41b8-8897-8004d4adb9da-kube-api-access-4vk28\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.547970 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkvf7\" (UniqueName: \"kubernetes.io/projected/1adeed8a-d531-46ee-b037-ea137468e026-kube-api-access-xkvf7\") pod \"ironic-operator-controller-manager-6f787dddc9-cb6dz\" (UID: \"1adeed8a-d531-46ee-b037-ea137468e026\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.554920 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.557890 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.571630 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.577407 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.577468 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.581893 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.582587 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-tbs6z" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.582763 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-xk4qs" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.586501 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.587611 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.590268 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-v49jp" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.604486 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.605870 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvsws\" (UniqueName: \"kubernetes.io/projected/71693924-c907-41df-b4ab-cbd9bfb7f97d-kube-api-access-hvsws\") pod \"manila-operator-controller-manager-55f864c847-sv6f4\" (UID: \"71693924-c907-41df-b4ab-cbd9bfb7f97d\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.606080 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72rhm\" (UniqueName: \"kubernetes.io/projected/1435162c-2ff0-4013-b838-48b63a57933b-kube-api-access-72rhm\") pod \"mariadb-operator-controller-manager-67ccfc9778-4pprb\" (UID: \"1435162c-2ff0-4013-b838-48b63a57933b\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.606129 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldh7q\" (UniqueName: \"kubernetes.io/projected/3dfa5be0-6028-4091-8129-71fa07ab93ac-kube-api-access-ldh7q\") pod \"keystone-operator-controller-manager-768b96df4c-bl6x4\" (UID: \"3dfa5be0-6028-4091-8129-71fa07ab93ac\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.612409 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.623489 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.631438 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.645023 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldh7q\" (UniqueName: \"kubernetes.io/projected/3dfa5be0-6028-4091-8129-71fa07ab93ac-kube-api-access-ldh7q\") pod \"keystone-operator-controller-manager-768b96df4c-bl6x4\" (UID: \"3dfa5be0-6028-4091-8129-71fa07ab93ac\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.666402 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.697158 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvsws\" (UniqueName: \"kubernetes.io/projected/71693924-c907-41df-b4ab-cbd9bfb7f97d-kube-api-access-hvsws\") pod \"manila-operator-controller-manager-55f864c847-sv6f4\" (UID: \"71693924-c907-41df-b4ab-cbd9bfb7f97d\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.705478 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.719044 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89lgd\" (UniqueName: \"kubernetes.io/projected/44f13e51-a8b3-489d-a4bb-852a441759f8-kube-api-access-89lgd\") pod \"neutron-operator-controller-manager-767865f676-n4k5l\" (UID: \"44f13e51-a8b3-489d-a4bb-852a441759f8\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.733467 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72rhm\" (UniqueName: \"kubernetes.io/projected/1435162c-2ff0-4013-b838-48b63a57933b-kube-api-access-72rhm\") pod \"mariadb-operator-controller-manager-67ccfc9778-4pprb\" (UID: \"1435162c-2ff0-4013-b838-48b63a57933b\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.733586 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-587f4\" (UniqueName: \"kubernetes.io/projected/8aa33b96-bd71-46c2-814a-ade6d5181f8a-kube-api-access-587f4\") pod \"octavia-operator-controller-manager-5b9f45d989-tw6tg\" (UID: \"8aa33b96-bd71-46c2-814a-ade6d5181f8a\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.733732 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr84r\" (UniqueName: \"kubernetes.io/projected/213a0cad-a19e-4337-ab03-67e0cc63fa08-kube-api-access-gr84r\") pod \"nova-operator-controller-manager-5d488d59fb-sxlpj\" (UID: \"213a0cad-a19e-4337-ab03-67e0cc63fa08\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.738970 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.740230 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.745639 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.748297 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-rmlbw" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.748469 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.754903 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72rhm\" (UniqueName: \"kubernetes.io/projected/1435162c-2ff0-4013-b838-48b63a57933b-kube-api-access-72rhm\") pod \"mariadb-operator-controller-manager-67ccfc9778-4pprb\" (UID: \"1435162c-2ff0-4013-b838-48b63a57933b\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.756104 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-tdb2g" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.756378 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.790361 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.792088 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.793010 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.800155 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-kvg89"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.801388 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.804588 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-z6xnx" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.809375 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-kvg89"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.824980 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.830981 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.835658 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89lgd\" (UniqueName: \"kubernetes.io/projected/44f13e51-a8b3-489d-a4bb-852a441759f8-kube-api-access-89lgd\") pod \"neutron-operator-controller-manager-767865f676-n4k5l\" (UID: \"44f13e51-a8b3-489d-a4bb-852a441759f8\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.835732 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-587f4\" (UniqueName: \"kubernetes.io/projected/8aa33b96-bd71-46c2-814a-ade6d5181f8a-kube-api-access-587f4\") pod \"octavia-operator-controller-manager-5b9f45d989-tw6tg\" (UID: \"8aa33b96-bd71-46c2-814a-ade6d5181f8a\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.835808 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn8vr\" (UniqueName: \"kubernetes.io/projected/87d69191-f1ca-46ea-a082-fdc249e342e1-kube-api-access-hn8vr\") pod \"ovn-operator-controller-manager-884679f54-cwdpq\" (UID: \"87d69191-f1ca-46ea-a082-fdc249e342e1\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.835843 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr84r\" (UniqueName: \"kubernetes.io/projected/213a0cad-a19e-4337-ab03-67e0cc63fa08-kube-api-access-gr84r\") pod \"nova-operator-controller-manager-5d488d59fb-sxlpj\" (UID: \"213a0cad-a19e-4337-ab03-67e0cc63fa08\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.837481 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.839621 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-76mcq" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.864298 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89lgd\" (UniqueName: \"kubernetes.io/projected/44f13e51-a8b3-489d-a4bb-852a441759f8-kube-api-access-89lgd\") pod \"neutron-operator-controller-manager-767865f676-n4k5l\" (UID: \"44f13e51-a8b3-489d-a4bb-852a441759f8\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.864770 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr84r\" (UniqueName: \"kubernetes.io/projected/213a0cad-a19e-4337-ab03-67e0cc63fa08-kube-api-access-gr84r\") pod \"nova-operator-controller-manager-5d488d59fb-sxlpj\" (UID: \"213a0cad-a19e-4337-ab03-67e0cc63fa08\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.867890 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-587f4\" (UniqueName: \"kubernetes.io/projected/8aa33b96-bd71-46c2-814a-ade6d5181f8a-kube-api-access-587f4\") pod \"octavia-operator-controller-manager-5b9f45d989-tw6tg\" (UID: \"8aa33b96-bd71-46c2-814a-ade6d5181f8a\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.868711 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.874522 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.876531 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.876946 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.881149 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-789bd" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.884454 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.894371 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.920377 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.924632 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.928128 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.929068 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bl64k" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.929090 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.937896 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn8vr\" (UniqueName: \"kubernetes.io/projected/87d69191-f1ca-46ea-a082-fdc249e342e1-kube-api-access-hn8vr\") pod \"ovn-operator-controller-manager-884679f54-cwdpq\" (UID: \"87d69191-f1ca-46ea-a082-fdc249e342e1\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.937980 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thfjp\" (UniqueName: \"kubernetes.io/projected/9e387ade-406a-4372-a097-554d1572296c-kube-api-access-thfjp\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.938071 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.938462 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq7bf\" (UniqueName: \"kubernetes.io/projected/eeb95c09-093f-48d0-8a80-d52fc8cb7157-kube-api-access-rq7bf\") pod \"placement-operator-controller-manager-5784578c99-kvg89\" (UID: \"eeb95c09-093f-48d0-8a80-d52fc8cb7157\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.938550 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mhcv\" (UniqueName: \"kubernetes.io/projected/ab86f223-d7e2-4527-9b1b-eb8633bafd01-kube-api-access-7mhcv\") pod \"swift-operator-controller-manager-c674c5965-rbl7w\" (UID: \"ab86f223-d7e2-4527-9b1b-eb8633bafd01\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.969803 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn8vr\" (UniqueName: \"kubernetes.io/projected/87d69191-f1ca-46ea-a082-fdc249e342e1-kube-api-access-hn8vr\") pod \"ovn-operator-controller-manager-884679f54-cwdpq\" (UID: \"87d69191-f1ca-46ea-a082-fdc249e342e1\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.980799 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2"] Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.981965 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.989021 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-j5kd4" Mar 20 09:16:08 crc kubenswrapper[4858]: I0320 09:16:08.998903 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2"] Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.039838 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mhcv\" (UniqueName: \"kubernetes.io/projected/ab86f223-d7e2-4527-9b1b-eb8633bafd01-kube-api-access-7mhcv\") pod \"swift-operator-controller-manager-c674c5965-rbl7w\" (UID: \"ab86f223-d7e2-4527-9b1b-eb8633bafd01\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.040435 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thfjp\" (UniqueName: \"kubernetes.io/projected/9e387ade-406a-4372-a097-554d1572296c-kube-api-access-thfjp\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.040482 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.040514 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.040554 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whknr\" (UniqueName: \"kubernetes.io/projected/a5c893bf-acfe-4ec9-a63a-3c48055530f4-kube-api-access-whknr\") pod \"telemetry-operator-controller-manager-d6b694c5-86cpb\" (UID: \"a5c893bf-acfe-4ec9-a63a-3c48055530f4\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.040578 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq7bf\" (UniqueName: \"kubernetes.io/projected/eeb95c09-093f-48d0-8a80-d52fc8cb7157-kube-api-access-rq7bf\") pod \"placement-operator-controller-manager-5784578c99-kvg89\" (UID: \"eeb95c09-093f-48d0-8a80-d52fc8cb7157\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.040609 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sbtk\" (UniqueName: \"kubernetes.io/projected/d4c562cc-1a1b-4c09-b468-314dd82774f3-kube-api-access-5sbtk\") pod \"test-operator-controller-manager-5c5cb9c4d7-zgtwl\" (UID: \"d4c562cc-1a1b-4c09-b468-314dd82774f3\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.041668 4858 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.041732 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert podName:9e387ade-406a-4372-a097-554d1572296c nodeName:}" failed. No retries permitted until 2026-03-20 09:16:09.541711934 +0000 UTC m=+1150.862130131 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" (UID: "9e387ade-406a-4372-a097-554d1572296c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.041882 4858 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.041925 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert podName:a0e45ec4-0059-41b8-8897-8004d4adb9da nodeName:}" failed. No retries permitted until 2026-03-20 09:16:10.041916099 +0000 UTC m=+1151.362334296 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert") pod "infra-operator-controller-manager-577ccd856-nw2q9" (UID: "a0e45ec4-0059-41b8-8897-8004d4adb9da") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.073426 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.076882 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj"] Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.077996 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.082880 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.083225 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-z4pn7" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.083394 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.092161 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mhcv\" (UniqueName: \"kubernetes.io/projected/ab86f223-d7e2-4527-9b1b-eb8633bafd01-kube-api-access-7mhcv\") pod \"swift-operator-controller-manager-c674c5965-rbl7w\" (UID: \"ab86f223-d7e2-4527-9b1b-eb8633bafd01\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.092950 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq7bf\" (UniqueName: \"kubernetes.io/projected/eeb95c09-093f-48d0-8a80-d52fc8cb7157-kube-api-access-rq7bf\") pod \"placement-operator-controller-manager-5784578c99-kvg89\" (UID: \"eeb95c09-093f-48d0-8a80-d52fc8cb7157\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.093450 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thfjp\" (UniqueName: \"kubernetes.io/projected/9e387ade-406a-4372-a097-554d1572296c-kube-api-access-thfjp\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.101304 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj"] Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.155747 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdr8m\" (UniqueName: \"kubernetes.io/projected/b91d223a-93ba-4db6-8153-9388e0e8a3a4-kube-api-access-jdr8m\") pod \"watcher-operator-controller-manager-6c4d75f7f9-zgbc2\" (UID: \"b91d223a-93ba-4db6-8153-9388e0e8a3a4\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.155826 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whknr\" (UniqueName: \"kubernetes.io/projected/a5c893bf-acfe-4ec9-a63a-3c48055530f4-kube-api-access-whknr\") pod \"telemetry-operator-controller-manager-d6b694c5-86cpb\" (UID: \"a5c893bf-acfe-4ec9-a63a-3c48055530f4\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.155867 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sbtk\" (UniqueName: \"kubernetes.io/projected/d4c562cc-1a1b-4c09-b468-314dd82774f3-kube-api-access-5sbtk\") pod \"test-operator-controller-manager-5c5cb9c4d7-zgtwl\" (UID: \"d4c562cc-1a1b-4c09-b468-314dd82774f3\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.246776 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whknr\" (UniqueName: \"kubernetes.io/projected/a5c893bf-acfe-4ec9-a63a-3c48055530f4-kube-api-access-whknr\") pod \"telemetry-operator-controller-manager-d6b694c5-86cpb\" (UID: \"a5c893bf-acfe-4ec9-a63a-3c48055530f4\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.249307 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.255182 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sbtk\" (UniqueName: \"kubernetes.io/projected/d4c562cc-1a1b-4c09-b468-314dd82774f3-kube-api-access-5sbtk\") pod \"test-operator-controller-manager-5c5cb9c4d7-zgtwl\" (UID: \"d4c562cc-1a1b-4c09-b468-314dd82774f3\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.263815 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm8fv\" (UniqueName: \"kubernetes.io/projected/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-kube-api-access-zm8fv\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.263998 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdr8m\" (UniqueName: \"kubernetes.io/projected/b91d223a-93ba-4db6-8153-9388e0e8a3a4-kube-api-access-jdr8m\") pod \"watcher-operator-controller-manager-6c4d75f7f9-zgbc2\" (UID: \"b91d223a-93ba-4db6-8153-9388e0e8a3a4\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.264025 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.265269 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.266575 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"3e965e3610cd2725487d92954a70f9b9ba52dc33b5e31fec564f254185703f58"} Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.307863 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.313959 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdr8m\" (UniqueName: \"kubernetes.io/projected/b91d223a-93ba-4db6-8153-9388e0e8a3a4-kube-api-access-jdr8m\") pod \"watcher-operator-controller-manager-6c4d75f7f9-zgbc2\" (UID: \"b91d223a-93ba-4db6-8153-9388e0e8a3a4\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.367421 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv"] Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.367817 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.367894 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm8fv\" (UniqueName: \"kubernetes.io/projected/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-kube-api-access-zm8fv\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.367967 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.368150 4858 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.368223 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:09.868198666 +0000 UTC m=+1151.188616863 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.368566 4858 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.368786 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:09.868753611 +0000 UTC m=+1151.189171998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "metrics-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.400202 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm8fv\" (UniqueName: \"kubernetes.io/projected/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-kube-api-access-zm8fv\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.410857 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4"] Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.500698 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.528751 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.554817 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.568760 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.574173 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.574449 4858 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.574599 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert podName:9e387ade-406a-4372-a097-554d1572296c nodeName:}" failed. No retries permitted until 2026-03-20 09:16:10.574558501 +0000 UTC m=+1151.894976698 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" (UID: "9e387ade-406a-4372-a097-554d1572296c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.830626 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb"] Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.838300 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf"] Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.881084 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: I0320 09:16:09.881166 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.881632 4858 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.881740 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:10.881711251 +0000 UTC m=+1152.202129448 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "metrics-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.882275 4858 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:16:09 crc kubenswrapper[4858]: E0320 09:16:09.882359 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:10.882338678 +0000 UTC m=+1152.202756875 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "webhook-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.087201 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.087423 4858 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.087472 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert podName:a0e45ec4-0059-41b8-8897-8004d4adb9da nodeName:}" failed. No retries permitted until 2026-03-20 09:16:12.08745691 +0000 UTC m=+1153.407875107 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert") pod "infra-operator-controller-manager-577ccd856-nw2q9" (UID: "a0e45ec4-0059-41b8-8897-8004d4adb9da") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.162929 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.233413 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4"] Mar 20 09:16:10 crc kubenswrapper[4858]: W0320 09:16:10.241393 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1adeed8a_d531_46ee_b037_ea137468e026.slice/crio-8d96005d2e903a49c4a726ba8375aa62d48a862575fbd15bac342346bb1142ee WatchSource:0}: Error finding container 8d96005d2e903a49c4a726ba8375aa62d48a862575fbd15bac342346bb1142ee: Status 404 returned error can't find the container with id 8d96005d2e903a49c4a726ba8375aa62d48a862575fbd15bac342346bb1142ee Mar 20 09:16:10 crc kubenswrapper[4858]: W0320 09:16:10.242200 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dfa5be0_6028_4091_8129_71fa07ab93ac.slice/crio-af600b619a6f9e1aeffc91027c90568aae644dbaf46370e4157c49e0e9a170b8 WatchSource:0}: Error finding container af600b619a6f9e1aeffc91027c90568aae644dbaf46370e4157c49e0e9a170b8: Status 404 returned error can't find the container with id af600b619a6f9e1aeffc91027c90568aae644dbaf46370e4157c49e0e9a170b8 Mar 20 09:16:10 crc kubenswrapper[4858]: W0320 09:16:10.248370 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e41ad93_0360_4a67_a155_ee05abfbcee1.slice/crio-696b6b62f376b8df824bcd991e88e9478f7b2aed686c0f6eeb8edaafa2220682 WatchSource:0}: Error finding container 696b6b62f376b8df824bcd991e88e9478f7b2aed686c0f6eeb8edaafa2220682: Status 404 returned error can't find the container with id 696b6b62f376b8df824bcd991e88e9478f7b2aed686c0f6eeb8edaafa2220682 Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.252821 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.257260 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.264228 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.274767 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" event={"ID":"0e2d5479-7826-4759-99ed-c3775a01035c","Type":"ContainerStarted","Data":"040f0d90ba5e5edf8987259ce35c4e0a3d67d1ecd978d18e4f6fba84fb7489dd"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.282920 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" event={"ID":"81033dc9-91dc-4bff-b69d-e1171f82b83c","Type":"ContainerStarted","Data":"0c748a067f08fecc26818754c2fa5ddd6e0993e64bd778f78dcc3b0d06e943cd"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.286543 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" event={"ID":"cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e","Type":"ContainerStarted","Data":"438f1acf79255c5c49dc6d7bc83e068ec0c2277e6685f43b41d538d90cabc0a2"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.287997 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" event={"ID":"1e41ad93-0360-4a67-a155-ee05abfbcee1","Type":"ContainerStarted","Data":"696b6b62f376b8df824bcd991e88e9478f7b2aed686c0f6eeb8edaafa2220682"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.289416 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" event={"ID":"71693924-c907-41df-b4ab-cbd9bfb7f97d","Type":"ContainerStarted","Data":"d3976cae96e521200b0d64003eb45dfe6fcb0151a65a9b82c7c56d8269071278"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.290711 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" event={"ID":"1adeed8a-d531-46ee-b037-ea137468e026","Type":"ContainerStarted","Data":"8d96005d2e903a49c4a726ba8375aa62d48a862575fbd15bac342346bb1142ee"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.292747 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" event={"ID":"3dfa5be0-6028-4091-8129-71fa07ab93ac","Type":"ContainerStarted","Data":"af600b619a6f9e1aeffc91027c90568aae644dbaf46370e4157c49e0e9a170b8"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.294539 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" event={"ID":"e2a119a6-367e-4a85-8617-e83d9f6a23b7","Type":"ContainerStarted","Data":"061c21384346f5e2061087d99c8013ecb7c918ac16c39dc570ed4b7019d4dd88"} Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.413445 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.432426 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.438784 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l"] Mar 20 09:16:10 crc kubenswrapper[4858]: W0320 09:16:10.444048 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfeae143f_10cc_4412_9c35_70499e77b5bd.slice/crio-35538db585322c3793dbc12c48c19b4a00c9a3fc9a842cf2691ff5297d0b5b71 WatchSource:0}: Error finding container 35538db585322c3793dbc12c48c19b4a00c9a3fc9a842cf2691ff5297d0b5b71: Status 404 returned error can't find the container with id 35538db585322c3793dbc12c48c19b4a00c9a3fc9a842cf2691ff5297d0b5b71 Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.474931 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.495006 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.506067 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w"] Mar 20 09:16:10 crc kubenswrapper[4858]: W0320 09:16:10.534252 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podab86f223_d7e2_4527_9b1b_eb8633bafd01.slice/crio-9a03408b25f653c2cab9c6c22c5b469c04576e42bc8745866f061a912f17d8d7 WatchSource:0}: Error finding container 9a03408b25f653c2cab9c6c22c5b469c04576e42bc8745866f061a912f17d8d7: Status 404 returned error can't find the container with id 9a03408b25f653c2cab9c6c22c5b469c04576e42bc8745866f061a912f17d8d7 Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.574994 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.583841 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-kvg89"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.590811 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb"] Mar 20 09:16:10 crc kubenswrapper[4858]: W0320 09:16:10.591151 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb91d223a_93ba_4db6_8153_9388e0e8a3a4.slice/crio-2e520bf38d7b669775a8dc7ceff2eb4bfeb22e86b52c6cd091a55339fd8d0559 WatchSource:0}: Error finding container 2e520bf38d7b669775a8dc7ceff2eb4bfeb22e86b52c6cd091a55339fd8d0559: Status 404 returned error can't find the container with id 2e520bf38d7b669775a8dc7ceff2eb4bfeb22e86b52c6cd091a55339fd8d0559 Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.594125 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-whknr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d6b694c5-86cpb_openstack-operators(a5c893bf-acfe-4ec9-a63a-3c48055530f4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.595656 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rq7bf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-kvg89_openstack-operators(eeb95c09-093f-48d0-8a80-d52fc8cb7157): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.595725 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jdr8m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-zgbc2_openstack-operators(b91d223a-93ba-4db6-8153-9388e0e8a3a4): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.595407 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" podUID="a5c893bf-acfe-4ec9-a63a-3c48055530f4" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.596775 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" podUID="eeb95c09-093f-48d0-8a80-d52fc8cb7157" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.596838 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" podUID="b91d223a-93ba-4db6-8153-9388e0e8a3a4" Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.596949 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl"] Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.599912 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.600150 4858 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.600218 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert podName:9e387ade-406a-4372-a097-554d1572296c nodeName:}" failed. No retries permitted until 2026-03-20 09:16:12.600194094 +0000 UTC m=+1153.920612291 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" (UID: "9e387ade-406a-4372-a097-554d1572296c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.617683 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5sbtk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-zgtwl_openstack-operators(d4c562cc-1a1b-4c09-b468-314dd82774f3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.618780 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" podUID="d4c562cc-1a1b-4c09-b468-314dd82774f3" Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.909502 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:10 crc kubenswrapper[4858]: I0320 09:16:10.909624 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.909806 4858 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.909858 4858 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.909923 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:12.909895191 +0000 UTC m=+1154.230313388 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "metrics-server-cert" not found Mar 20 09:16:10 crc kubenswrapper[4858]: E0320 09:16:10.909949 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:12.909940873 +0000 UTC m=+1154.230359070 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "webhook-server-cert" not found Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.321164 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" event={"ID":"213a0cad-a19e-4337-ab03-67e0cc63fa08","Type":"ContainerStarted","Data":"2a0d7ed4d9ffbaa75c6defac74a0e4f97c3d1d7c5022024f2e09a118184be795"} Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.323822 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" event={"ID":"44f13e51-a8b3-489d-a4bb-852a441759f8","Type":"ContainerStarted","Data":"55ec731da151aade2784df7a0e0b560547492d784aedc584533adf7240f8fe1f"} Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.326280 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" event={"ID":"87d69191-f1ca-46ea-a082-fdc249e342e1","Type":"ContainerStarted","Data":"90a896868bd5b6a115ebdc693b702c1345cd0fd73193462f32fbaec6ab9c6b5d"} Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.328351 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" event={"ID":"1435162c-2ff0-4013-b838-48b63a57933b","Type":"ContainerStarted","Data":"04a03863b329f4bfbff59ec3f26cef5590b13c8ae1e36087d89b2efb9579a779"} Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.329845 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" event={"ID":"ab86f223-d7e2-4527-9b1b-eb8633bafd01","Type":"ContainerStarted","Data":"9a03408b25f653c2cab9c6c22c5b469c04576e42bc8745866f061a912f17d8d7"} Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.332869 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" event={"ID":"eeb95c09-093f-48d0-8a80-d52fc8cb7157","Type":"ContainerStarted","Data":"16914ea9b011c95bd69ce89b86c50166c520b2e3814e8cdb52af07d8d852cff8"} Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.335787 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" event={"ID":"8aa33b96-bd71-46c2-814a-ade6d5181f8a","Type":"ContainerStarted","Data":"4a5c615ca266ded8d1b483b1ac79f75c6df409374e7e17ca9a1eb7cffdd9ca87"} Mar 20 09:16:11 crc kubenswrapper[4858]: E0320 09:16:11.338822 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" podUID="eeb95c09-093f-48d0-8a80-d52fc8cb7157" Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.342798 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" event={"ID":"a5c893bf-acfe-4ec9-a63a-3c48055530f4","Type":"ContainerStarted","Data":"1e92ad7c8c0dd6a155220f49108a5a34e7d8ac45f64ec5137da55149b9408e0a"} Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.345496 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" event={"ID":"feae143f-10cc-4412-9c35-70499e77b5bd","Type":"ContainerStarted","Data":"35538db585322c3793dbc12c48c19b4a00c9a3fc9a842cf2691ff5297d0b5b71"} Mar 20 09:16:11 crc kubenswrapper[4858]: E0320 09:16:11.347563 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" podUID="a5c893bf-acfe-4ec9-a63a-3c48055530f4" Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.363795 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" event={"ID":"d4c562cc-1a1b-4c09-b468-314dd82774f3","Type":"ContainerStarted","Data":"c530e7e4b318c1a12120cac18e276c665f5800e6d25bea3b31bb2ac96e22eb3e"} Mar 20 09:16:11 crc kubenswrapper[4858]: E0320 09:16:11.365345 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" podUID="d4c562cc-1a1b-4c09-b468-314dd82774f3" Mar 20 09:16:11 crc kubenswrapper[4858]: I0320 09:16:11.366065 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" event={"ID":"b91d223a-93ba-4db6-8153-9388e0e8a3a4","Type":"ContainerStarted","Data":"2e520bf38d7b669775a8dc7ceff2eb4bfeb22e86b52c6cd091a55339fd8d0559"} Mar 20 09:16:11 crc kubenswrapper[4858]: E0320 09:16:11.366965 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" podUID="b91d223a-93ba-4db6-8153-9388e0e8a3a4" Mar 20 09:16:12 crc kubenswrapper[4858]: I0320 09:16:12.138638 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.138883 4858 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.138955 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert podName:a0e45ec4-0059-41b8-8897-8004d4adb9da nodeName:}" failed. No retries permitted until 2026-03-20 09:16:16.13891752 +0000 UTC m=+1157.459335707 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert") pod "infra-operator-controller-manager-577ccd856-nw2q9" (UID: "a0e45ec4-0059-41b8-8897-8004d4adb9da") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.377741 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" podUID="d4c562cc-1a1b-4c09-b468-314dd82774f3" Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.380142 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" podUID="a5c893bf-acfe-4ec9-a63a-3c48055530f4" Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.380200 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" podUID="b91d223a-93ba-4db6-8153-9388e0e8a3a4" Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.380553 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" podUID="eeb95c09-093f-48d0-8a80-d52fc8cb7157" Mar 20 09:16:12 crc kubenswrapper[4858]: I0320 09:16:12.646272 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.646650 4858 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.646748 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert podName:9e387ade-406a-4372-a097-554d1572296c nodeName:}" failed. No retries permitted until 2026-03-20 09:16:16.646720063 +0000 UTC m=+1157.967138270 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" (UID: "9e387ade-406a-4372-a097-554d1572296c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:12 crc kubenswrapper[4858]: I0320 09:16:12.955093 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:12 crc kubenswrapper[4858]: I0320 09:16:12.955229 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.955339 4858 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.955448 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:16.955421554 +0000 UTC m=+1158.275839801 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "webhook-server-cert" not found Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.955499 4858 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:16:12 crc kubenswrapper[4858]: E0320 09:16:12.955633 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:16.955601409 +0000 UTC m=+1158.276019606 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "metrics-server-cert" not found Mar 20 09:16:16 crc kubenswrapper[4858]: I0320 09:16:16.217857 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:16 crc kubenswrapper[4858]: E0320 09:16:16.218849 4858 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:16 crc kubenswrapper[4858]: E0320 09:16:16.218898 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert podName:a0e45ec4-0059-41b8-8897-8004d4adb9da nodeName:}" failed. No retries permitted until 2026-03-20 09:16:24.218881111 +0000 UTC m=+1165.539299308 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert") pod "infra-operator-controller-manager-577ccd856-nw2q9" (UID: "a0e45ec4-0059-41b8-8897-8004d4adb9da") : secret "infra-operator-webhook-server-cert" not found Mar 20 09:16:16 crc kubenswrapper[4858]: I0320 09:16:16.727332 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:16 crc kubenswrapper[4858]: E0320 09:16:16.727518 4858 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:16 crc kubenswrapper[4858]: E0320 09:16:16.727612 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert podName:9e387ade-406a-4372-a097-554d1572296c nodeName:}" failed. No retries permitted until 2026-03-20 09:16:24.727588108 +0000 UTC m=+1166.048006305 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" (UID: "9e387ade-406a-4372-a097-554d1572296c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:17 crc kubenswrapper[4858]: I0320 09:16:17.032212 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:17 crc kubenswrapper[4858]: I0320 09:16:17.032288 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:17 crc kubenswrapper[4858]: E0320 09:16:17.032620 4858 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:16:17 crc kubenswrapper[4858]: E0320 09:16:17.032792 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:25.032752165 +0000 UTC m=+1166.353170532 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "webhook-server-cert" not found Mar 20 09:16:17 crc kubenswrapper[4858]: E0320 09:16:17.033190 4858 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:16:17 crc kubenswrapper[4858]: E0320 09:16:17.033413 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:25.033389622 +0000 UTC m=+1166.353807839 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "metrics-server-cert" not found Mar 20 09:16:22 crc kubenswrapper[4858]: E0320 09:16:22.554535 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad" Mar 20 09:16:22 crc kubenswrapper[4858]: E0320 09:16:22.555538 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-86rzd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-588d4d986b-8bwmb_openstack-operators(e2a119a6-367e-4a85-8617-e83d9f6a23b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:22 crc kubenswrapper[4858]: E0320 09:16:22.556755 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" podUID="e2a119a6-367e-4a85-8617-e83d9f6a23b7" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.291197 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.291481 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xkvf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6f787dddc9-cb6dz_openstack-operators(1adeed8a-d531-46ee-b037-ea137468e026): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.292725 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" podUID="1adeed8a-d531-46ee-b037-ea137468e026" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.468750 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:9dd26bc51e7757d84736528d4988a1f980ad50ccb070aef6fc252e32c5c423a8\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" podUID="1adeed8a-d531-46ee-b037-ea137468e026" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.468768 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:12841b27173f5f1beeb83112e057c8753f4cf411f583fba4f0610fac0f60b7ad\\\"\"" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" podUID="e2a119a6-367e-4a85-8617-e83d9f6a23b7" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.921064 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.921296 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v97hw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-79df6bcc97-bb9mf_openstack-operators(81033dc9-91dc-4bff-b69d-e1171f82b83c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:23 crc kubenswrapper[4858]: E0320 09:16:23.922496 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" podUID="81033dc9-91dc-4bff-b69d-e1171f82b83c" Mar 20 09:16:24 crc kubenswrapper[4858]: I0320 09:16:24.267228 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:24 crc kubenswrapper[4858]: I0320 09:16:24.274171 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0e45ec4-0059-41b8-8897-8004d4adb9da-cert\") pod \"infra-operator-controller-manager-577ccd856-nw2q9\" (UID: \"a0e45ec4-0059-41b8-8897-8004d4adb9da\") " pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:24 crc kubenswrapper[4858]: I0320 09:16:24.308421 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:24 crc kubenswrapper[4858]: E0320 09:16:24.477075 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:76a1cde9f29fb39ed715b06be16adb803b9a2e24d68acb369911c0a88e33bc7d\\\"\"" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" podUID="81033dc9-91dc-4bff-b69d-e1171f82b83c" Mar 20 09:16:24 crc kubenswrapper[4858]: E0320 09:16:24.554173 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da" Mar 20 09:16:24 crc kubenswrapper[4858]: E0320 09:16:24.554375 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hvsws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-55f864c847-sv6f4_openstack-operators(71693924-c907-41df-b4ab-cbd9bfb7f97d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:24 crc kubenswrapper[4858]: E0320 09:16:24.555686 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" podUID="71693924-c907-41df-b4ab-cbd9bfb7f97d" Mar 20 09:16:24 crc kubenswrapper[4858]: I0320 09:16:24.775179 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:24 crc kubenswrapper[4858]: E0320 09:16:24.775458 4858 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:24 crc kubenswrapper[4858]: E0320 09:16:24.775653 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert podName:9e387ade-406a-4372-a097-554d1572296c nodeName:}" failed. No retries permitted until 2026-03-20 09:16:40.775620604 +0000 UTC m=+1182.096038811 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" (UID: "9e387ade-406a-4372-a097-554d1572296c") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.071221 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55" Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.071494 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hn8vr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-884679f54-cwdpq_openstack-operators(87d69191-f1ca-46ea-a082-fdc249e342e1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.073329 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" podUID="87d69191-f1ca-46ea-a082-fdc249e342e1" Mar 20 09:16:25 crc kubenswrapper[4858]: I0320 09:16:25.080422 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:25 crc kubenswrapper[4858]: I0320 09:16:25.080616 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.080657 4858 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.080764 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:41.080730604 +0000 UTC m=+1182.401148981 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "webhook-server-cert" not found Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.080796 4858 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.080881 4858 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs podName:4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7 nodeName:}" failed. No retries permitted until 2026-03-20 09:16:41.080856007 +0000 UTC m=+1182.401274194 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs") pod "openstack-operator-controller-manager-55958644c4-zzhvj" (UID: "4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7") : secret "metrics-server-cert" not found Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.487019 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:bef93f71d3b42a72d8b96c69bdb4db4b8bd797c5093a0a719443d7a5c9aaab55\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" podUID="87d69191-f1ca-46ea-a082-fdc249e342e1" Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.487074 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f2e0b0fb34995b8acbbf1b0b60b5dbcf488b4f3899d1bb0763ae7dcee9bae6da\\\"\"" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" podUID="71693924-c907-41df-b4ab-cbd9bfb7f97d" Mar 20 09:16:25 crc kubenswrapper[4858]: I0320 09:16:25.753444 4858 scope.go:117] "RemoveContainer" containerID="006663afa8696a0b72fbe29cee51b6e695e4265a339622aaebf05e1b3da841e6" Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.968606 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56" Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.968897 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ldh7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-768b96df4c-bl6x4_openstack-operators(3dfa5be0-6028-4091-8129-71fa07ab93ac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:25 crc kubenswrapper[4858]: E0320 09:16:25.970088 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" podUID="3dfa5be0-6028-4091-8129-71fa07ab93ac" Mar 20 09:16:26 crc kubenswrapper[4858]: E0320 09:16:26.494134 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:ec36a9083657587022f8471c9d5a71b87a7895398496e7fc546c73aa1eae4b56\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" podUID="3dfa5be0-6028-4091-8129-71fa07ab93ac" Mar 20 09:16:26 crc kubenswrapper[4858]: E0320 09:16:26.991686 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a" Mar 20 09:16:26 crc kubenswrapper[4858]: E0320 09:16:26.991859 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gr84r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5d488d59fb-sxlpj_openstack-operators(213a0cad-a19e-4337-ab03-67e0cc63fa08): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:26 crc kubenswrapper[4858]: E0320 09:16:26.993239 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" podUID="213a0cad-a19e-4337-ab03-67e0cc63fa08" Mar 20 09:16:27 crc kubenswrapper[4858]: E0320 09:16:27.502022 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" podUID="213a0cad-a19e-4337-ab03-67e0cc63fa08" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.481676 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9"] Mar 20 09:16:30 crc kubenswrapper[4858]: W0320 09:16:30.491014 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0e45ec4_0059_41b8_8897_8004d4adb9da.slice/crio-90342931e00df1d11f706e63661be7249f77468e9da673243047f863d0382566 WatchSource:0}: Error finding container 90342931e00df1d11f706e63661be7249f77468e9da673243047f863d0382566: Status 404 returned error can't find the container with id 90342931e00df1d11f706e63661be7249f77468e9da673243047f863d0382566 Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.532877 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" event={"ID":"ab86f223-d7e2-4527-9b1b-eb8633bafd01","Type":"ContainerStarted","Data":"23c485a226f4d4fec489418b9d0e548938d6dc85e48434ba67b6813e7051a2ee"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.534393 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.539242 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" event={"ID":"b91d223a-93ba-4db6-8153-9388e0e8a3a4","Type":"ContainerStarted","Data":"5f22f68c98a5d45613cecb835472ac456aa6a38a7a7a2d20741990a35b2deb0a"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.539832 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.544554 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" event={"ID":"feae143f-10cc-4412-9c35-70499e77b5bd","Type":"ContainerStarted","Data":"c37a618098d903eac0f9ac892b4e6057e997ea2cce815c783035db917af034c9"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.544783 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.552974 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" event={"ID":"0e2d5479-7826-4759-99ed-c3775a01035c","Type":"ContainerStarted","Data":"ddbc5063301d7fe964ab431086f37afc03956da1c56c1e31b34e0ae137796d2a"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.553694 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.560582 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" podStartSLOduration=5.399754007 podStartE2EDuration="22.560564883s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.545133724 +0000 UTC m=+1151.865551921" lastFinishedPulling="2026-03-20 09:16:27.7059446 +0000 UTC m=+1169.026362797" observedRunningTime="2026-03-20 09:16:30.557651724 +0000 UTC m=+1171.878069921" watchObservedRunningTime="2026-03-20 09:16:30.560564883 +0000 UTC m=+1171.880983080" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.561638 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" event={"ID":"cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e","Type":"ContainerStarted","Data":"26c81f76215a419e768b83a247c6a9cb51495b76a349345dfaed79246b475317"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.563789 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.575118 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" event={"ID":"a0e45ec4-0059-41b8-8897-8004d4adb9da","Type":"ContainerStarted","Data":"90342931e00df1d11f706e63661be7249f77468e9da673243047f863d0382566"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.581597 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.584249 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" event={"ID":"1435162c-2ff0-4013-b838-48b63a57933b","Type":"ContainerStarted","Data":"71b8af424da0ed3219160d27df4ea96deb03a327c3feec9175748f1a20a2b2b1"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.585012 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.594356 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" podStartSLOduration=4.749604972 podStartE2EDuration="22.59433876s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.455083804 +0000 UTC m=+1151.775502001" lastFinishedPulling="2026-03-20 09:16:28.299817552 +0000 UTC m=+1169.620235789" observedRunningTime="2026-03-20 09:16:30.591920536 +0000 UTC m=+1171.912338753" watchObservedRunningTime="2026-03-20 09:16:30.59433876 +0000 UTC m=+1171.914756977" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.602889 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" event={"ID":"1e41ad93-0360-4a67-a155-ee05abfbcee1","Type":"ContainerStarted","Data":"d62653959440e78509149b54441ffc2c9c274386179bc43b556dc73b2d38b92c"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.603586 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.608506 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" event={"ID":"8aa33b96-bd71-46c2-814a-ade6d5181f8a","Type":"ContainerStarted","Data":"0cb194bdb66ec124ebfac45555e79730016ca5cacc533c5f58d7cf2a6f0a64b7"} Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.609372 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.630836 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" podStartSLOduration=6.354038429 podStartE2EDuration="23.63081081s" podCreationTimestamp="2026-03-20 09:16:07 +0000 UTC" firstStartedPulling="2026-03-20 09:16:09.429976835 +0000 UTC m=+1150.750395032" lastFinishedPulling="2026-03-20 09:16:26.706749186 +0000 UTC m=+1168.027167413" observedRunningTime="2026-03-20 09:16:30.618157201 +0000 UTC m=+1171.938575418" watchObservedRunningTime="2026-03-20 09:16:30.63081081 +0000 UTC m=+1171.951229007" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.649597 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" podStartSLOduration=3.199192565 podStartE2EDuration="22.649573735s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.595528871 +0000 UTC m=+1151.915947068" lastFinishedPulling="2026-03-20 09:16:30.045910031 +0000 UTC m=+1171.366328238" observedRunningTime="2026-03-20 09:16:30.645137466 +0000 UTC m=+1171.965555673" watchObservedRunningTime="2026-03-20 09:16:30.649573735 +0000 UTC m=+1171.969991942" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.767921 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" podStartSLOduration=6.433755832 podStartE2EDuration="23.767903455s" podCreationTimestamp="2026-03-20 09:16:07 +0000 UTC" firstStartedPulling="2026-03-20 09:16:09.372598473 +0000 UTC m=+1150.693016670" lastFinishedPulling="2026-03-20 09:16:26.706746096 +0000 UTC m=+1168.027164293" observedRunningTime="2026-03-20 09:16:30.678003249 +0000 UTC m=+1171.998421446" watchObservedRunningTime="2026-03-20 09:16:30.767903455 +0000 UTC m=+1172.088321862" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.824460 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" podStartSLOduration=5.379200051 podStartE2EDuration="22.824431485s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.261242481 +0000 UTC m=+1151.581660678" lastFinishedPulling="2026-03-20 09:16:27.706473915 +0000 UTC m=+1169.026892112" observedRunningTime="2026-03-20 09:16:30.767572966 +0000 UTC m=+1172.087991163" watchObservedRunningTime="2026-03-20 09:16:30.824431485 +0000 UTC m=+1172.144849682" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.859577 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" podStartSLOduration=5.607237573 podStartE2EDuration="22.859555999s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.454147249 +0000 UTC m=+1151.774565446" lastFinishedPulling="2026-03-20 09:16:27.706465675 +0000 UTC m=+1169.026883872" observedRunningTime="2026-03-20 09:16:30.859500057 +0000 UTC m=+1172.179918264" watchObservedRunningTime="2026-03-20 09:16:30.859555999 +0000 UTC m=+1172.179974196" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.860739 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" podStartSLOduration=6.407979816 podStartE2EDuration="22.86073262s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.27292253 +0000 UTC m=+1151.593340727" lastFinishedPulling="2026-03-20 09:16:26.725675324 +0000 UTC m=+1168.046093531" observedRunningTime="2026-03-20 09:16:30.826706385 +0000 UTC m=+1172.147124582" watchObservedRunningTime="2026-03-20 09:16:30.86073262 +0000 UTC m=+1172.181150817" Mar 20 09:16:30 crc kubenswrapper[4858]: I0320 09:16:30.973850 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" podStartSLOduration=5.1301058489999996 podStartE2EDuration="22.97382243s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.454636502 +0000 UTC m=+1151.775054699" lastFinishedPulling="2026-03-20 09:16:28.298353043 +0000 UTC m=+1169.618771280" observedRunningTime="2026-03-20 09:16:30.943777092 +0000 UTC m=+1172.264195299" watchObservedRunningTime="2026-03-20 09:16:30.97382243 +0000 UTC m=+1172.294240627" Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.619469 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" event={"ID":"a5c893bf-acfe-4ec9-a63a-3c48055530f4","Type":"ContainerStarted","Data":"4ce6440b235260b74a96cfbed2014ac42cb0c3c25fcf111d0a7918ac1aea6ae3"} Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.619803 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.624257 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" event={"ID":"44f13e51-a8b3-489d-a4bb-852a441759f8","Type":"ContainerStarted","Data":"f0752f765bc05af66adf051066d7eb0498aa61a3a0020cbd3679ab584b1ace7e"} Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.626091 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" event={"ID":"d4c562cc-1a1b-4c09-b468-314dd82774f3","Type":"ContainerStarted","Data":"0f0c61057543d06314eb362afbac01d76ca0d2df101581f7bcac6b7eedc3fa79"} Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.626368 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.629593 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" event={"ID":"eeb95c09-093f-48d0-8a80-d52fc8cb7157","Type":"ContainerStarted","Data":"2de44507caefddce7edbbe92cedf79997271fe7e598dd87bcd6c7826a6b23e2a"} Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.648240 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" podStartSLOduration=4.201064612 podStartE2EDuration="23.648212545s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.593819535 +0000 UTC m=+1151.914237732" lastFinishedPulling="2026-03-20 09:16:30.040967468 +0000 UTC m=+1171.361385665" observedRunningTime="2026-03-20 09:16:31.642748559 +0000 UTC m=+1172.963166776" watchObservedRunningTime="2026-03-20 09:16:31.648212545 +0000 UTC m=+1172.968630742" Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.669588 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" podStartSLOduration=4.241949621 podStartE2EDuration="23.669561109s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.595498119 +0000 UTC m=+1151.915916316" lastFinishedPulling="2026-03-20 09:16:30.023109607 +0000 UTC m=+1171.343527804" observedRunningTime="2026-03-20 09:16:31.66404315 +0000 UTC m=+1172.984461367" watchObservedRunningTime="2026-03-20 09:16:31.669561109 +0000 UTC m=+1172.989979306" Mar 20 09:16:31 crc kubenswrapper[4858]: I0320 09:16:31.698871 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" podStartSLOduration=4.260481371 podStartE2EDuration="23.698843026s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.617512693 +0000 UTC m=+1151.937930890" lastFinishedPulling="2026-03-20 09:16:30.055874348 +0000 UTC m=+1171.376292545" observedRunningTime="2026-03-20 09:16:31.692081854 +0000 UTC m=+1173.012500061" watchObservedRunningTime="2026-03-20 09:16:31.698843026 +0000 UTC m=+1173.019261223" Mar 20 09:16:33 crc kubenswrapper[4858]: I0320 09:16:33.645989 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" event={"ID":"a0e45ec4-0059-41b8-8897-8004d4adb9da","Type":"ContainerStarted","Data":"65c169d8660fd34ee5a56d4e31dbfe867324e4d415ec5aef91898f47c353b981"} Mar 20 09:16:33 crc kubenswrapper[4858]: I0320 09:16:33.646484 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:33 crc kubenswrapper[4858]: I0320 09:16:33.662711 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" podStartSLOduration=22.938972672 podStartE2EDuration="25.662677487s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:30.49872065 +0000 UTC m=+1171.819138847" lastFinishedPulling="2026-03-20 09:16:33.222425465 +0000 UTC m=+1174.542843662" observedRunningTime="2026-03-20 09:16:33.659838981 +0000 UTC m=+1174.980257198" watchObservedRunningTime="2026-03-20 09:16:33.662677487 +0000 UTC m=+1174.983095714" Mar 20 09:16:37 crc kubenswrapper[4858]: I0320 09:16:37.688639 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" event={"ID":"1adeed8a-d531-46ee-b037-ea137468e026","Type":"ContainerStarted","Data":"48db39d9e2583d743aea29eb51b450455564d54c685b2f66bfa4c2e6ad3289a1"} Mar 20 09:16:37 crc kubenswrapper[4858]: I0320 09:16:37.689841 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" Mar 20 09:16:37 crc kubenswrapper[4858]: I0320 09:16:37.710272 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" podStartSLOduration=3.34097013 podStartE2EDuration="29.710248193s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.260111441 +0000 UTC m=+1151.580529638" lastFinishedPulling="2026-03-20 09:16:36.629389504 +0000 UTC m=+1177.949807701" observedRunningTime="2026-03-20 09:16:37.705815464 +0000 UTC m=+1179.026233651" watchObservedRunningTime="2026-03-20 09:16:37.710248193 +0000 UTC m=+1179.030666390" Mar 20 09:16:38 crc kubenswrapper[4858]: I0320 09:16:38.399541 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-7c7lv" Mar 20 09:16:38 crc kubenswrapper[4858]: I0320 09:16:38.415743 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-zk2d4" Mar 20 09:16:38 crc kubenswrapper[4858]: I0320 09:16:38.586606 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-ng56m" Mar 20 09:16:38 crc kubenswrapper[4858]: I0320 09:16:38.607139 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-sk6sh" Mar 20 09:16:38 crc kubenswrapper[4858]: I0320 09:16:38.872235 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-tw6tg" Mar 20 09:16:38 crc kubenswrapper[4858]: I0320 09:16:38.879413 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-4pprb" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.076241 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-n4k5l" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.308784 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.312077 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-kvg89" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.506460 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-rbl7w" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.535066 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-86cpb" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.558704 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-zgtwl" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.589643 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-zgbc2" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.753659 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" event={"ID":"81033dc9-91dc-4bff-b69d-e1171f82b83c","Type":"ContainerStarted","Data":"3e32bad26fbaa8671255c7837e445470da030c94adbdb0f3fcceaad6120d696e"} Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.754676 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.770717 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" event={"ID":"87d69191-f1ca-46ea-a082-fdc249e342e1","Type":"ContainerStarted","Data":"55c59c56792966a7d9c73687d7ce727ae2f11d551927b1f5ef82f05b2dc2a4d3"} Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.771463 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.778285 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" event={"ID":"3dfa5be0-6028-4091-8129-71fa07ab93ac","Type":"ContainerStarted","Data":"f45db22b3a73db8b5b20dbb41d5de4c2ec74ef7ab0b27dd042973fb777b63a16"} Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.779142 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.790032 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" event={"ID":"e2a119a6-367e-4a85-8617-e83d9f6a23b7","Type":"ContainerStarted","Data":"0c982c87fa2c97be6ed60097df986f0cb5222099b8e57beeca39e338b66a1291"} Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.790395 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.791053 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" podStartSLOduration=3.052202552 podStartE2EDuration="31.791030826s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:09.89863361 +0000 UTC m=+1151.219051807" lastFinishedPulling="2026-03-20 09:16:38.637461884 +0000 UTC m=+1179.957880081" observedRunningTime="2026-03-20 09:16:39.77512678 +0000 UTC m=+1181.095544987" watchObservedRunningTime="2026-03-20 09:16:39.791030826 +0000 UTC m=+1181.111449013" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.815791 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" podStartSLOduration=3.578603618 podStartE2EDuration="31.815773951s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.489499117 +0000 UTC m=+1151.809917314" lastFinishedPulling="2026-03-20 09:16:38.72666944 +0000 UTC m=+1180.047087647" observedRunningTime="2026-03-20 09:16:39.812260967 +0000 UTC m=+1181.132679164" watchObservedRunningTime="2026-03-20 09:16:39.815773951 +0000 UTC m=+1181.136192148" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.845023 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" podStartSLOduration=3.588188736 podStartE2EDuration="31.844999338s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.257475071 +0000 UTC m=+1151.577893268" lastFinishedPulling="2026-03-20 09:16:38.514285663 +0000 UTC m=+1179.834703870" observedRunningTime="2026-03-20 09:16:39.836986192 +0000 UTC m=+1181.157404389" watchObservedRunningTime="2026-03-20 09:16:39.844999338 +0000 UTC m=+1181.165417535" Mar 20 09:16:39 crc kubenswrapper[4858]: I0320 09:16:39.862411 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" podStartSLOduration=3.196575592 podStartE2EDuration="31.862389945s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:09.90505272 +0000 UTC m=+1151.225470917" lastFinishedPulling="2026-03-20 09:16:38.570867073 +0000 UTC m=+1179.891285270" observedRunningTime="2026-03-20 09:16:39.862096787 +0000 UTC m=+1181.182514984" watchObservedRunningTime="2026-03-20 09:16:39.862389945 +0000 UTC m=+1181.182808162" Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.798922 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" event={"ID":"71693924-c907-41df-b4ab-cbd9bfb7f97d","Type":"ContainerStarted","Data":"e0f887b382c0fe134d46e9c43c3b3afc742345dce0d107cc8f622ded9c9cbbe7"} Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.799224 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.801675 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" event={"ID":"213a0cad-a19e-4337-ab03-67e0cc63fa08","Type":"ContainerStarted","Data":"76d2b8c5ff4a6372a018119052263d8cee7353aee1f011137ca750882e6c45e0"} Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.802306 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.803817 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.817960 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9e387ade-406a-4372-a097-554d1572296c-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-9gjsl\" (UID: \"9e387ade-406a-4372-a097-554d1572296c\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.821306 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" podStartSLOduration=2.360903961 podStartE2EDuration="32.821277736s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.170589696 +0000 UTC m=+1151.491007893" lastFinishedPulling="2026-03-20 09:16:40.630963471 +0000 UTC m=+1181.951381668" observedRunningTime="2026-03-20 09:16:40.817440803 +0000 UTC m=+1182.137859000" watchObservedRunningTime="2026-03-20 09:16:40.821277736 +0000 UTC m=+1182.141695953" Mar 20 09:16:40 crc kubenswrapper[4858]: I0320 09:16:40.838212 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" podStartSLOduration=3.729690729 podStartE2EDuration="32.83819072s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:10.461616108 +0000 UTC m=+1151.782034305" lastFinishedPulling="2026-03-20 09:16:39.570116099 +0000 UTC m=+1180.890534296" observedRunningTime="2026-03-20 09:16:40.834155222 +0000 UTC m=+1182.154573429" watchObservedRunningTime="2026-03-20 09:16:40.83819072 +0000 UTC m=+1182.158608917" Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.093615 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.116054 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.116197 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.121913 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-webhook-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.123943 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7-metrics-certs\") pod \"openstack-operator-controller-manager-55958644c4-zzhvj\" (UID: \"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7\") " pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.396911 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.566468 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl"] Mar 20 09:16:41 crc kubenswrapper[4858]: W0320 09:16:41.571620 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e387ade_406a_4372_a097_554d1572296c.slice/crio-18b7efb871166398d65fce78c70420c6fb520245433eeee9a0fc6d7f3e317612 WatchSource:0}: Error finding container 18b7efb871166398d65fce78c70420c6fb520245433eeee9a0fc6d7f3e317612: Status 404 returned error can't find the container with id 18b7efb871166398d65fce78c70420c6fb520245433eeee9a0fc6d7f3e317612 Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.809931 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" event={"ID":"9e387ade-406a-4372-a097-554d1572296c","Type":"ContainerStarted","Data":"18b7efb871166398d65fce78c70420c6fb520245433eeee9a0fc6d7f3e317612"} Mar 20 09:16:41 crc kubenswrapper[4858]: I0320 09:16:41.896307 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj"] Mar 20 09:16:42 crc kubenswrapper[4858]: I0320 09:16:42.821274 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" event={"ID":"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7","Type":"ContainerStarted","Data":"fcdc3a0cabbea5417acf5a38c6b66c005fa79286c12ea157520a163198aa0c9f"} Mar 20 09:16:42 crc kubenswrapper[4858]: I0320 09:16:42.821756 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" event={"ID":"4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7","Type":"ContainerStarted","Data":"cfc8c211b177129eed5c019061aae07a4fcc20b4a3e62fdaf1d96c82d8549d9c"} Mar 20 09:16:42 crc kubenswrapper[4858]: I0320 09:16:42.822249 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:42 crc kubenswrapper[4858]: I0320 09:16:42.868041 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" podStartSLOduration=34.868013225 podStartE2EDuration="34.868013225s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-20 09:16:42.861684515 +0000 UTC m=+1184.182102722" watchObservedRunningTime="2026-03-20 09:16:42.868013225 +0000 UTC m=+1184.188431432" Mar 20 09:16:44 crc kubenswrapper[4858]: I0320 09:16:44.315253 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-577ccd856-nw2q9" Mar 20 09:16:48 crc kubenswrapper[4858]: I0320 09:16:48.535104 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-8bwmb" Mar 20 09:16:48 crc kubenswrapper[4858]: I0320 09:16:48.561424 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-bb9mf" Mar 20 09:16:48 crc kubenswrapper[4858]: I0320 09:16:48.635928 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-cb6dz" Mar 20 09:16:48 crc kubenswrapper[4858]: I0320 09:16:48.797067 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-sv6f4" Mar 20 09:16:48 crc kubenswrapper[4858]: I0320 09:16:48.830418 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-bl6x4" Mar 20 09:16:48 crc kubenswrapper[4858]: I0320 09:16:48.934349 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-sxlpj" Mar 20 09:16:49 crc kubenswrapper[4858]: I0320 09:16:49.252025 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-cwdpq" Mar 20 09:16:51 crc kubenswrapper[4858]: I0320 09:16:51.404075 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-55958644c4-zzhvj" Mar 20 09:16:52 crc kubenswrapper[4858]: E0320 09:16:52.955465 4858 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329" Mar 20 09:16:52 crc kubenswrapper[4858]: E0320 09:16:52.955986 4858 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thfjp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-89d64c458-9gjsl_openstack-operators(9e387ade-406a-4372-a097-554d1572296c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 20 09:16:52 crc kubenswrapper[4858]: E0320 09:16:52.957347 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" podUID="9e387ade-406a-4372-a097-554d1572296c" Mar 20 09:16:53 crc kubenswrapper[4858]: E0320 09:16:53.924558 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bf7cdbfb125c4327b35870f8640cbed9ddc32d6f07fedd117c6fd59f16463329\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" podUID="9e387ade-406a-4372-a097-554d1572296c" Mar 20 09:17:06 crc kubenswrapper[4858]: I0320 09:17:06.074235 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:17:07 crc kubenswrapper[4858]: I0320 09:17:07.033695 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" event={"ID":"9e387ade-406a-4372-a097-554d1572296c","Type":"ContainerStarted","Data":"6d2c21cdb8b1001dc24286b67dbfcf1a991e5e81ebe57782fda48c4577981bd7"} Mar 20 09:17:07 crc kubenswrapper[4858]: I0320 09:17:07.034624 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:17:07 crc kubenswrapper[4858]: I0320 09:17:07.072284 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" podStartSLOduration=33.9793019 podStartE2EDuration="59.072262229s" podCreationTimestamp="2026-03-20 09:16:08 +0000 UTC" firstStartedPulling="2026-03-20 09:16:41.589340169 +0000 UTC m=+1182.909758366" lastFinishedPulling="2026-03-20 09:17:06.682300508 +0000 UTC m=+1208.002718695" observedRunningTime="2026-03-20 09:17:07.066563446 +0000 UTC m=+1208.386981663" watchObservedRunningTime="2026-03-20 09:17:07.072262229 +0000 UTC m=+1208.392680426" Mar 20 09:17:11 crc kubenswrapper[4858]: I0320 09:17:11.106968 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-9gjsl" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.167254 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-cmrz2"] Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.170293 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.176971 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.177677 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.177712 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.178166 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-98dtm" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.182679 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-cmrz2"] Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.194118 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sp5x\" (UniqueName: \"kubernetes.io/projected/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-kube-api-access-8sp5x\") pod \"dnsmasq-dns-675f4bcbfc-cmrz2\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.194338 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-config\") pod \"dnsmasq-dns-675f4bcbfc-cmrz2\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.261290 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tgjql"] Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.265155 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.272967 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.278770 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tgjql"] Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.296640 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55debadc-f9be-4dc0-a269-4e8782024065-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.296732 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sp5x\" (UniqueName: \"kubernetes.io/projected/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-kube-api-access-8sp5x\") pod \"dnsmasq-dns-675f4bcbfc-cmrz2\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.296782 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-config\") pod \"dnsmasq-dns-675f4bcbfc-cmrz2\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.296803 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55debadc-f9be-4dc0-a269-4e8782024065-config\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.296866 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlkt7\" (UniqueName: \"kubernetes.io/projected/55debadc-f9be-4dc0-a269-4e8782024065-kube-api-access-qlkt7\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.298295 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-config\") pod \"dnsmasq-dns-675f4bcbfc-cmrz2\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.336702 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sp5x\" (UniqueName: \"kubernetes.io/projected/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-kube-api-access-8sp5x\") pod \"dnsmasq-dns-675f4bcbfc-cmrz2\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.398920 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlkt7\" (UniqueName: \"kubernetes.io/projected/55debadc-f9be-4dc0-a269-4e8782024065-kube-api-access-qlkt7\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.399385 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55debadc-f9be-4dc0-a269-4e8782024065-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.399455 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55debadc-f9be-4dc0-a269-4e8782024065-config\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.400402 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55debadc-f9be-4dc0-a269-4e8782024065-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.400417 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55debadc-f9be-4dc0-a269-4e8782024065-config\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.426431 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlkt7\" (UniqueName: \"kubernetes.io/projected/55debadc-f9be-4dc0-a269-4e8782024065-kube-api-access-qlkt7\") pod \"dnsmasq-dns-78dd6ddcc-tgjql\" (UID: \"55debadc-f9be-4dc0-a269-4e8782024065\") " pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.503123 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:25 crc kubenswrapper[4858]: I0320 09:17:25.587095 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:26 crc kubenswrapper[4858]: I0320 09:17:26.227600 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-cmrz2"] Mar 20 09:17:26 crc kubenswrapper[4858]: I0320 09:17:26.242126 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-tgjql"] Mar 20 09:17:27 crc kubenswrapper[4858]: I0320 09:17:27.227369 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" event={"ID":"55debadc-f9be-4dc0-a269-4e8782024065","Type":"ContainerStarted","Data":"44e743f771da14aebde84f29cccfa9548e7fd71594430c0d435c73f4e328db8f"} Mar 20 09:17:27 crc kubenswrapper[4858]: I0320 09:17:27.230345 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" event={"ID":"c890ea46-ed6c-42eb-a995-621ab7cd8e2e","Type":"ContainerStarted","Data":"9fa89e4df4e0875425dcef4d7690af01b2609a3a16e5a988400fbec9e9832777"} Mar 20 09:17:39 crc kubenswrapper[4858]: I0320 09:17:39.346723 4858 generic.go:334] "Generic (PLEG): container finished" podID="55debadc-f9be-4dc0-a269-4e8782024065" containerID="864ebefe0b7c8768bb83711e5bf8db33f74a7f00c8f82926c0b89b08235e815b" exitCode=0 Mar 20 09:17:39 crc kubenswrapper[4858]: I0320 09:17:39.346792 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" event={"ID":"55debadc-f9be-4dc0-a269-4e8782024065","Type":"ContainerDied","Data":"864ebefe0b7c8768bb83711e5bf8db33f74a7f00c8f82926c0b89b08235e815b"} Mar 20 09:17:39 crc kubenswrapper[4858]: I0320 09:17:39.350942 4858 generic.go:334] "Generic (PLEG): container finished" podID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerID="70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962" exitCode=0 Mar 20 09:17:39 crc kubenswrapper[4858]: I0320 09:17:39.351004 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" event={"ID":"c890ea46-ed6c-42eb-a995-621ab7cd8e2e","Type":"ContainerDied","Data":"70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962"} Mar 20 09:17:40 crc kubenswrapper[4858]: I0320 09:17:40.362803 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" event={"ID":"c890ea46-ed6c-42eb-a995-621ab7cd8e2e","Type":"ContainerStarted","Data":"a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871"} Mar 20 09:17:40 crc kubenswrapper[4858]: I0320 09:17:40.363129 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:40 crc kubenswrapper[4858]: I0320 09:17:40.366940 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" event={"ID":"55debadc-f9be-4dc0-a269-4e8782024065","Type":"ContainerStarted","Data":"247ac4c47eb5ac3cc3ad1280e701ef14b7a82c93052668de16301fa929c7e206"} Mar 20 09:17:40 crc kubenswrapper[4858]: I0320 09:17:40.367121 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:40 crc kubenswrapper[4858]: I0320 09:17:40.392775 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" podStartSLOduration=3.149835023 podStartE2EDuration="15.392743736s" podCreationTimestamp="2026-03-20 09:17:25 +0000 UTC" firstStartedPulling="2026-03-20 09:17:26.233034574 +0000 UTC m=+1227.553452771" lastFinishedPulling="2026-03-20 09:17:38.475943297 +0000 UTC m=+1239.796361484" observedRunningTime="2026-03-20 09:17:40.384270084 +0000 UTC m=+1241.704688291" watchObservedRunningTime="2026-03-20 09:17:40.392743736 +0000 UTC m=+1241.713161943" Mar 20 09:17:40 crc kubenswrapper[4858]: I0320 09:17:40.415592 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" podStartSLOduration=3.225770337 podStartE2EDuration="15.415562423s" podCreationTimestamp="2026-03-20 09:17:25 +0000 UTC" firstStartedPulling="2026-03-20 09:17:26.249616068 +0000 UTC m=+1227.570034265" lastFinishedPulling="2026-03-20 09:17:38.439408154 +0000 UTC m=+1239.759826351" observedRunningTime="2026-03-20 09:17:40.407544453 +0000 UTC m=+1241.727962660" watchObservedRunningTime="2026-03-20 09:17:40.415562423 +0000 UTC m=+1241.735980630" Mar 20 09:17:45 crc kubenswrapper[4858]: I0320 09:17:45.504671 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:45 crc kubenswrapper[4858]: I0320 09:17:45.589574 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78dd6ddcc-tgjql" Mar 20 09:17:45 crc kubenswrapper[4858]: I0320 09:17:45.640912 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-cmrz2"] Mar 20 09:17:46 crc kubenswrapper[4858]: I0320 09:17:46.415378 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" podUID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerName="dnsmasq-dns" containerID="cri-o://a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871" gracePeriod=10 Mar 20 09:17:46 crc kubenswrapper[4858]: I0320 09:17:46.891659 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.031011 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-config\") pod \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.031094 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sp5x\" (UniqueName: \"kubernetes.io/projected/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-kube-api-access-8sp5x\") pod \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\" (UID: \"c890ea46-ed6c-42eb-a995-621ab7cd8e2e\") " Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.040111 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-kube-api-access-8sp5x" (OuterVolumeSpecName: "kube-api-access-8sp5x") pod "c890ea46-ed6c-42eb-a995-621ab7cd8e2e" (UID: "c890ea46-ed6c-42eb-a995-621ab7cd8e2e"). InnerVolumeSpecName "kube-api-access-8sp5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.070196 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-config" (OuterVolumeSpecName: "config") pod "c890ea46-ed6c-42eb-a995-621ab7cd8e2e" (UID: "c890ea46-ed6c-42eb-a995-621ab7cd8e2e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.133247 4858 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-config\") on node \"crc\" DevicePath \"\"" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.133298 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sp5x\" (UniqueName: \"kubernetes.io/projected/c890ea46-ed6c-42eb-a995-621ab7cd8e2e-kube-api-access-8sp5x\") on node \"crc\" DevicePath \"\"" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.426922 4858 generic.go:334] "Generic (PLEG): container finished" podID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerID="a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871" exitCode=0 Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.427018 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.427023 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" event={"ID":"c890ea46-ed6c-42eb-a995-621ab7cd8e2e","Type":"ContainerDied","Data":"a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871"} Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.430224 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-cmrz2" event={"ID":"c890ea46-ed6c-42eb-a995-621ab7cd8e2e","Type":"ContainerDied","Data":"9fa89e4df4e0875425dcef4d7690af01b2609a3a16e5a988400fbec9e9832777"} Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.430256 4858 scope.go:117] "RemoveContainer" containerID="a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.457572 4858 scope.go:117] "RemoveContainer" containerID="70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.463865 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-cmrz2"] Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.469684 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-cmrz2"] Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.495767 4858 scope.go:117] "RemoveContainer" containerID="a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871" Mar 20 09:17:47 crc kubenswrapper[4858]: E0320 09:17:47.496499 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871\": container with ID starting with a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871 not found: ID does not exist" containerID="a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.496563 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871"} err="failed to get container status \"a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871\": rpc error: code = NotFound desc = could not find container \"a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871\": container with ID starting with a86764d47109e62982137ef2f4c17667d8cd31d3a2f3efca10929333a2884871 not found: ID does not exist" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.496603 4858 scope.go:117] "RemoveContainer" containerID="70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962" Mar 20 09:17:47 crc kubenswrapper[4858]: E0320 09:17:47.497141 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962\": container with ID starting with 70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962 not found: ID does not exist" containerID="70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962" Mar 20 09:17:47 crc kubenswrapper[4858]: I0320 09:17:47.497203 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962"} err="failed to get container status \"70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962\": rpc error: code = NotFound desc = could not find container \"70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962\": container with ID starting with 70fc81ad7c19c33edd9ab8be0844f4a07cf4e93e7cb79791af4c3c44e10ba962 not found: ID does not exist" Mar 20 09:17:48 crc kubenswrapper[4858]: I0320 09:17:48.082540 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" path="/var/lib/kubelet/pods/c890ea46-ed6c-42eb-a995-621ab7cd8e2e/volumes" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.155368 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566638-nr9hv"] Mar 20 09:18:00 crc kubenswrapper[4858]: E0320 09:18:00.156798 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerName="dnsmasq-dns" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.156818 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerName="dnsmasq-dns" Mar 20 09:18:00 crc kubenswrapper[4858]: E0320 09:18:00.156846 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerName="init" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.156852 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerName="init" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.156998 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="c890ea46-ed6c-42eb-a995-621ab7cd8e2e" containerName="dnsmasq-dns" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.157615 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-nr9hv" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.164577 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.164995 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.165165 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.173492 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-nr9hv"] Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.278466 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj2kz\" (UniqueName: \"kubernetes.io/projected/9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc-kube-api-access-nj2kz\") pod \"auto-csr-approver-29566638-nr9hv\" (UID: \"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc\") " pod="openshift-infra/auto-csr-approver-29566638-nr9hv" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.380488 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj2kz\" (UniqueName: \"kubernetes.io/projected/9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc-kube-api-access-nj2kz\") pod \"auto-csr-approver-29566638-nr9hv\" (UID: \"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc\") " pod="openshift-infra/auto-csr-approver-29566638-nr9hv" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.406630 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj2kz\" (UniqueName: \"kubernetes.io/projected/9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc-kube-api-access-nj2kz\") pod \"auto-csr-approver-29566638-nr9hv\" (UID: \"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc\") " pod="openshift-infra/auto-csr-approver-29566638-nr9hv" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.483464 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-nr9hv" Mar 20 09:18:00 crc kubenswrapper[4858]: I0320 09:18:00.982524 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-nr9hv"] Mar 20 09:18:01 crc kubenswrapper[4858]: I0320 09:18:01.591734 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566638-nr9hv" event={"ID":"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc","Type":"ContainerStarted","Data":"75205efb345b232b9b740969874ab8c636d4786908e25c804fa59e52f28b5bc0"} Mar 20 09:18:05 crc kubenswrapper[4858]: I0320 09:18:05.633230 4858 generic.go:334] "Generic (PLEG): container finished" podID="9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc" containerID="8d4668e2a92f6c867a304e6f5397d6cf565dc1d8110e4e18a0772adad7914d90" exitCode=0 Mar 20 09:18:05 crc kubenswrapper[4858]: I0320 09:18:05.633355 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566638-nr9hv" event={"ID":"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc","Type":"ContainerDied","Data":"8d4668e2a92f6c867a304e6f5397d6cf565dc1d8110e4e18a0772adad7914d90"} Mar 20 09:18:06 crc kubenswrapper[4858]: I0320 09:18:06.938541 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-nr9hv" Mar 20 09:18:07 crc kubenswrapper[4858]: I0320 09:18:07.104940 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nj2kz\" (UniqueName: \"kubernetes.io/projected/9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc-kube-api-access-nj2kz\") pod \"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc\" (UID: \"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc\") " Mar 20 09:18:07 crc kubenswrapper[4858]: I0320 09:18:07.116780 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc-kube-api-access-nj2kz" (OuterVolumeSpecName: "kube-api-access-nj2kz") pod "9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc" (UID: "9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc"). InnerVolumeSpecName "kube-api-access-nj2kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:18:07 crc kubenswrapper[4858]: I0320 09:18:07.207470 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nj2kz\" (UniqueName: \"kubernetes.io/projected/9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc-kube-api-access-nj2kz\") on node \"crc\" DevicePath \"\"" Mar 20 09:18:07 crc kubenswrapper[4858]: I0320 09:18:07.655370 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566638-nr9hv" event={"ID":"9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc","Type":"ContainerDied","Data":"75205efb345b232b9b740969874ab8c636d4786908e25c804fa59e52f28b5bc0"} Mar 20 09:18:07 crc kubenswrapper[4858]: I0320 09:18:07.655845 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75205efb345b232b9b740969874ab8c636d4786908e25c804fa59e52f28b5bc0" Mar 20 09:18:07 crc kubenswrapper[4858]: I0320 09:18:07.655450 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566638-nr9hv" Mar 20 09:18:08 crc kubenswrapper[4858]: I0320 09:18:08.039712 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-v974z"] Mar 20 09:18:08 crc kubenswrapper[4858]: I0320 09:18:08.046274 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566632-v974z"] Mar 20 09:18:08 crc kubenswrapper[4858]: I0320 09:18:08.080191 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="116c8acc-4ac8-4989-b2f2-70ff2f159f0f" path="/var/lib/kubelet/pods/116c8acc-4ac8-4989-b2f2-70ff2f159f0f/volumes" Mar 20 09:18:30 crc kubenswrapper[4858]: I0320 09:18:30.051910 4858 scope.go:117] "RemoveContainer" containerID="afb0eaaff81e64a736ed219c95e12e0066ce9b1793557e369f41ba31fefb0942" Mar 20 09:18:37 crc kubenswrapper[4858]: I0320 09:18:37.890935 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:18:37 crc kubenswrapper[4858]: I0320 09:18:37.891963 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:19:07 crc kubenswrapper[4858]: I0320 09:19:07.890127 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:19:07 crc kubenswrapper[4858]: I0320 09:19:07.891102 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:19:37 crc kubenswrapper[4858]: I0320 09:19:37.890200 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:19:37 crc kubenswrapper[4858]: I0320 09:19:37.891180 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:19:37 crc kubenswrapper[4858]: I0320 09:19:37.891247 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:19:37 crc kubenswrapper[4858]: I0320 09:19:37.892188 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3e965e3610cd2725487d92954a70f9b9ba52dc33b5e31fec564f254185703f58"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:19:37 crc kubenswrapper[4858]: I0320 09:19:37.892247 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://3e965e3610cd2725487d92954a70f9b9ba52dc33b5e31fec564f254185703f58" gracePeriod=600 Mar 20 09:19:38 crc kubenswrapper[4858]: I0320 09:19:38.418035 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="3e965e3610cd2725487d92954a70f9b9ba52dc33b5e31fec564f254185703f58" exitCode=0 Mar 20 09:19:38 crc kubenswrapper[4858]: I0320 09:19:38.418089 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"3e965e3610cd2725487d92954a70f9b9ba52dc33b5e31fec564f254185703f58"} Mar 20 09:19:38 crc kubenswrapper[4858]: I0320 09:19:38.418701 4858 scope.go:117] "RemoveContainer" containerID="bef7d78bb90262eb2557357139a02f6a23b1e0a616279703c46e019d97babf79" Mar 20 09:19:39 crc kubenswrapper[4858]: I0320 09:19:39.429698 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"1214de77b7ba6f5035904ddcf50d5dc1a7d89457797a1d34a3dfae1bc23c2fd0"} Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.150780 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566640-vpmsj"] Mar 20 09:20:00 crc kubenswrapper[4858]: E0320 09:20:00.152022 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc" containerName="oc" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.152082 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc" containerName="oc" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.152275 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc" containerName="oc" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.152889 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-vpmsj" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.154839 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.154922 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.156870 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.170920 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-vpmsj"] Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.252914 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wjtg\" (UniqueName: \"kubernetes.io/projected/ba8e3af1-a5eb-4855-a9c1-903713dac8c2-kube-api-access-9wjtg\") pod \"auto-csr-approver-29566640-vpmsj\" (UID: \"ba8e3af1-a5eb-4855-a9c1-903713dac8c2\") " pod="openshift-infra/auto-csr-approver-29566640-vpmsj" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.354773 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wjtg\" (UniqueName: \"kubernetes.io/projected/ba8e3af1-a5eb-4855-a9c1-903713dac8c2-kube-api-access-9wjtg\") pod \"auto-csr-approver-29566640-vpmsj\" (UID: \"ba8e3af1-a5eb-4855-a9c1-903713dac8c2\") " pod="openshift-infra/auto-csr-approver-29566640-vpmsj" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.391204 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wjtg\" (UniqueName: \"kubernetes.io/projected/ba8e3af1-a5eb-4855-a9c1-903713dac8c2-kube-api-access-9wjtg\") pod \"auto-csr-approver-29566640-vpmsj\" (UID: \"ba8e3af1-a5eb-4855-a9c1-903713dac8c2\") " pod="openshift-infra/auto-csr-approver-29566640-vpmsj" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.476109 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-vpmsj" Mar 20 09:20:00 crc kubenswrapper[4858]: I0320 09:20:00.912920 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-vpmsj"] Mar 20 09:20:01 crc kubenswrapper[4858]: I0320 09:20:01.616061 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566640-vpmsj" event={"ID":"ba8e3af1-a5eb-4855-a9c1-903713dac8c2","Type":"ContainerStarted","Data":"4ebf221bcf3478bfce2f7f390383b8233a047c3072a3d4666a4d82959007aa06"} Mar 20 09:20:03 crc kubenswrapper[4858]: I0320 09:20:03.633695 4858 generic.go:334] "Generic (PLEG): container finished" podID="ba8e3af1-a5eb-4855-a9c1-903713dac8c2" containerID="4a94e79ae5b566aacad8414fae2ea0b962213b680ca0078c4a63944122d88e95" exitCode=0 Mar 20 09:20:03 crc kubenswrapper[4858]: I0320 09:20:03.633786 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566640-vpmsj" event={"ID":"ba8e3af1-a5eb-4855-a9c1-903713dac8c2","Type":"ContainerDied","Data":"4a94e79ae5b566aacad8414fae2ea0b962213b680ca0078c4a63944122d88e95"} Mar 20 09:20:04 crc kubenswrapper[4858]: I0320 09:20:04.942180 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-vpmsj" Mar 20 09:20:05 crc kubenswrapper[4858]: I0320 09:20:05.032345 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wjtg\" (UniqueName: \"kubernetes.io/projected/ba8e3af1-a5eb-4855-a9c1-903713dac8c2-kube-api-access-9wjtg\") pod \"ba8e3af1-a5eb-4855-a9c1-903713dac8c2\" (UID: \"ba8e3af1-a5eb-4855-a9c1-903713dac8c2\") " Mar 20 09:20:05 crc kubenswrapper[4858]: I0320 09:20:05.040699 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba8e3af1-a5eb-4855-a9c1-903713dac8c2-kube-api-access-9wjtg" (OuterVolumeSpecName: "kube-api-access-9wjtg") pod "ba8e3af1-a5eb-4855-a9c1-903713dac8c2" (UID: "ba8e3af1-a5eb-4855-a9c1-903713dac8c2"). InnerVolumeSpecName "kube-api-access-9wjtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:20:05 crc kubenswrapper[4858]: I0320 09:20:05.135397 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wjtg\" (UniqueName: \"kubernetes.io/projected/ba8e3af1-a5eb-4855-a9c1-903713dac8c2-kube-api-access-9wjtg\") on node \"crc\" DevicePath \"\"" Mar 20 09:20:05 crc kubenswrapper[4858]: I0320 09:20:05.658515 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566640-vpmsj" event={"ID":"ba8e3af1-a5eb-4855-a9c1-903713dac8c2","Type":"ContainerDied","Data":"4ebf221bcf3478bfce2f7f390383b8233a047c3072a3d4666a4d82959007aa06"} Mar 20 09:20:05 crc kubenswrapper[4858]: I0320 09:20:05.658575 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ebf221bcf3478bfce2f7f390383b8233a047c3072a3d4666a4d82959007aa06" Mar 20 09:20:05 crc kubenswrapper[4858]: I0320 09:20:05.658642 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566640-vpmsj" Mar 20 09:20:06 crc kubenswrapper[4858]: I0320 09:20:06.025865 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-ppj7p"] Mar 20 09:20:06 crc kubenswrapper[4858]: I0320 09:20:06.031409 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566634-ppj7p"] Mar 20 09:20:06 crc kubenswrapper[4858]: I0320 09:20:06.081291 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="757ea99d-a6c3-4d69-a150-eb561f504a59" path="/var/lib/kubelet/pods/757ea99d-a6c3-4d69-a150-eb561f504a59/volumes" Mar 20 09:20:30 crc kubenswrapper[4858]: I0320 09:20:30.175748 4858 scope.go:117] "RemoveContainer" containerID="43a9e6bf47bed6b01e4b2601523433ff8ed6687fde7e09bdfd880700829ad0ee" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.143049 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566642-rlsxc"] Mar 20 09:22:00 crc kubenswrapper[4858]: E0320 09:22:00.146223 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba8e3af1-a5eb-4855-a9c1-903713dac8c2" containerName="oc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.146404 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba8e3af1-a5eb-4855-a9c1-903713dac8c2" containerName="oc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.146704 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba8e3af1-a5eb-4855-a9c1-903713dac8c2" containerName="oc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.147722 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-rlsxc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.153450 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.153835 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-rlsxc"] Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.154692 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.154925 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.307337 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcmsb\" (UniqueName: \"kubernetes.io/projected/adae7626-9eb0-46d9-b8d8-d5d0629181b4-kube-api-access-gcmsb\") pod \"auto-csr-approver-29566642-rlsxc\" (UID: \"adae7626-9eb0-46d9-b8d8-d5d0629181b4\") " pod="openshift-infra/auto-csr-approver-29566642-rlsxc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.408925 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcmsb\" (UniqueName: \"kubernetes.io/projected/adae7626-9eb0-46d9-b8d8-d5d0629181b4-kube-api-access-gcmsb\") pod \"auto-csr-approver-29566642-rlsxc\" (UID: \"adae7626-9eb0-46d9-b8d8-d5d0629181b4\") " pod="openshift-infra/auto-csr-approver-29566642-rlsxc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.429227 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcmsb\" (UniqueName: \"kubernetes.io/projected/adae7626-9eb0-46d9-b8d8-d5d0629181b4-kube-api-access-gcmsb\") pod \"auto-csr-approver-29566642-rlsxc\" (UID: \"adae7626-9eb0-46d9-b8d8-d5d0629181b4\") " pod="openshift-infra/auto-csr-approver-29566642-rlsxc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.491701 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-rlsxc" Mar 20 09:22:00 crc kubenswrapper[4858]: I0320 09:22:00.963381 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-rlsxc"] Mar 20 09:22:01 crc kubenswrapper[4858]: I0320 09:22:01.609364 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566642-rlsxc" event={"ID":"adae7626-9eb0-46d9-b8d8-d5d0629181b4","Type":"ContainerStarted","Data":"2f5e452f61c0b8fad7712facaad8f4472ae4c110f59fd7b403f4800e2b781a8a"} Mar 20 09:22:02 crc kubenswrapper[4858]: I0320 09:22:02.620057 4858 generic.go:334] "Generic (PLEG): container finished" podID="adae7626-9eb0-46d9-b8d8-d5d0629181b4" containerID="c76c5b41a4e11da041a4d4407fbad14b2e0953d23cc0c9a95d6449fe175bf58b" exitCode=0 Mar 20 09:22:02 crc kubenswrapper[4858]: I0320 09:22:02.620140 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566642-rlsxc" event={"ID":"adae7626-9eb0-46d9-b8d8-d5d0629181b4","Type":"ContainerDied","Data":"c76c5b41a4e11da041a4d4407fbad14b2e0953d23cc0c9a95d6449fe175bf58b"} Mar 20 09:22:03 crc kubenswrapper[4858]: I0320 09:22:03.906177 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-rlsxc" Mar 20 09:22:03 crc kubenswrapper[4858]: I0320 09:22:03.977723 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcmsb\" (UniqueName: \"kubernetes.io/projected/adae7626-9eb0-46d9-b8d8-d5d0629181b4-kube-api-access-gcmsb\") pod \"adae7626-9eb0-46d9-b8d8-d5d0629181b4\" (UID: \"adae7626-9eb0-46d9-b8d8-d5d0629181b4\") " Mar 20 09:22:03 crc kubenswrapper[4858]: I0320 09:22:03.985795 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adae7626-9eb0-46d9-b8d8-d5d0629181b4-kube-api-access-gcmsb" (OuterVolumeSpecName: "kube-api-access-gcmsb") pod "adae7626-9eb0-46d9-b8d8-d5d0629181b4" (UID: "adae7626-9eb0-46d9-b8d8-d5d0629181b4"). InnerVolumeSpecName "kube-api-access-gcmsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:22:04 crc kubenswrapper[4858]: I0320 09:22:04.080123 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcmsb\" (UniqueName: \"kubernetes.io/projected/adae7626-9eb0-46d9-b8d8-d5d0629181b4-kube-api-access-gcmsb\") on node \"crc\" DevicePath \"\"" Mar 20 09:22:04 crc kubenswrapper[4858]: I0320 09:22:04.636954 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566642-rlsxc" event={"ID":"adae7626-9eb0-46d9-b8d8-d5d0629181b4","Type":"ContainerDied","Data":"2f5e452f61c0b8fad7712facaad8f4472ae4c110f59fd7b403f4800e2b781a8a"} Mar 20 09:22:04 crc kubenswrapper[4858]: I0320 09:22:04.637001 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f5e452f61c0b8fad7712facaad8f4472ae4c110f59fd7b403f4800e2b781a8a" Mar 20 09:22:04 crc kubenswrapper[4858]: I0320 09:22:04.637536 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566642-rlsxc" Mar 20 09:22:04 crc kubenswrapper[4858]: I0320 09:22:04.988687 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-8b64v"] Mar 20 09:22:04 crc kubenswrapper[4858]: I0320 09:22:04.994209 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566636-8b64v"] Mar 20 09:22:06 crc kubenswrapper[4858]: I0320 09:22:06.082509 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1276f93-9455-48ff-82c1-3ddba162054f" path="/var/lib/kubelet/pods/d1276f93-9455-48ff-82c1-3ddba162054f/volumes" Mar 20 09:22:07 crc kubenswrapper[4858]: I0320 09:22:07.890981 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:22:07 crc kubenswrapper[4858]: I0320 09:22:07.891673 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:22:30 crc kubenswrapper[4858]: I0320 09:22:30.271637 4858 scope.go:117] "RemoveContainer" containerID="3de46f8786541609f7420ade58699a2ad681e40cebcc0404f7bd1ca364f3a1d8" Mar 20 09:22:37 crc kubenswrapper[4858]: I0320 09:22:37.890696 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:22:37 crc kubenswrapper[4858]: I0320 09:22:37.891487 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:23:07 crc kubenswrapper[4858]: I0320 09:23:07.890843 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:23:07 crc kubenswrapper[4858]: I0320 09:23:07.891779 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:23:07 crc kubenswrapper[4858]: I0320 09:23:07.891873 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:23:07 crc kubenswrapper[4858]: I0320 09:23:07.892931 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1214de77b7ba6f5035904ddcf50d5dc1a7d89457797a1d34a3dfae1bc23c2fd0"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:23:07 crc kubenswrapper[4858]: I0320 09:23:07.893017 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://1214de77b7ba6f5035904ddcf50d5dc1a7d89457797a1d34a3dfae1bc23c2fd0" gracePeriod=600 Mar 20 09:23:09 crc kubenswrapper[4858]: I0320 09:23:09.207579 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="1214de77b7ba6f5035904ddcf50d5dc1a7d89457797a1d34a3dfae1bc23c2fd0" exitCode=0 Mar 20 09:23:09 crc kubenswrapper[4858]: I0320 09:23:09.207720 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"1214de77b7ba6f5035904ddcf50d5dc1a7d89457797a1d34a3dfae1bc23c2fd0"} Mar 20 09:23:09 crc kubenswrapper[4858]: I0320 09:23:09.208517 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea"} Mar 20 09:23:09 crc kubenswrapper[4858]: I0320 09:23:09.208545 4858 scope.go:117] "RemoveContainer" containerID="3e965e3610cd2725487d92954a70f9b9ba52dc33b5e31fec564f254185703f58" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.143591 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hnb9t"] Mar 20 09:23:12 crc kubenswrapper[4858]: E0320 09:23:12.144093 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adae7626-9eb0-46d9-b8d8-d5d0629181b4" containerName="oc" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.144116 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="adae7626-9eb0-46d9-b8d8-d5d0629181b4" containerName="oc" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.144298 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="adae7626-9eb0-46d9-b8d8-d5d0629181b4" containerName="oc" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.145570 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.153142 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnb9t"] Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.340856 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-catalog-content\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.342138 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwtdc\" (UniqueName: \"kubernetes.io/projected/0afbf51b-beb7-4808-8002-39a67f948cd1-kube-api-access-fwtdc\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.342745 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-utilities\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.444397 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-utilities\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.444520 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-catalog-content\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.444580 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwtdc\" (UniqueName: \"kubernetes.io/projected/0afbf51b-beb7-4808-8002-39a67f948cd1-kube-api-access-fwtdc\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.445004 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-utilities\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.445163 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-catalog-content\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.483407 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwtdc\" (UniqueName: \"kubernetes.io/projected/0afbf51b-beb7-4808-8002-39a67f948cd1-kube-api-access-fwtdc\") pod \"redhat-operators-hnb9t\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:12 crc kubenswrapper[4858]: I0320 09:23:12.767993 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:13 crc kubenswrapper[4858]: I0320 09:23:13.040789 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnb9t"] Mar 20 09:23:13 crc kubenswrapper[4858]: I0320 09:23:13.248635 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnb9t" event={"ID":"0afbf51b-beb7-4808-8002-39a67f948cd1","Type":"ContainerStarted","Data":"7132aef89a55e7bb081c547cba02dbb98d2dd9b0ae7df9831f965b8465122d91"} Mar 20 09:23:14 crc kubenswrapper[4858]: I0320 09:23:14.258085 4858 generic.go:334] "Generic (PLEG): container finished" podID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerID="225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f" exitCode=0 Mar 20 09:23:14 crc kubenswrapper[4858]: I0320 09:23:14.258147 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnb9t" event={"ID":"0afbf51b-beb7-4808-8002-39a67f948cd1","Type":"ContainerDied","Data":"225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f"} Mar 20 09:23:14 crc kubenswrapper[4858]: I0320 09:23:14.260615 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:23:16 crc kubenswrapper[4858]: I0320 09:23:16.280766 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnb9t" event={"ID":"0afbf51b-beb7-4808-8002-39a67f948cd1","Type":"ContainerStarted","Data":"23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd"} Mar 20 09:23:17 crc kubenswrapper[4858]: I0320 09:23:17.292229 4858 generic.go:334] "Generic (PLEG): container finished" podID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerID="23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd" exitCode=0 Mar 20 09:23:17 crc kubenswrapper[4858]: I0320 09:23:17.292427 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnb9t" event={"ID":"0afbf51b-beb7-4808-8002-39a67f948cd1","Type":"ContainerDied","Data":"23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd"} Mar 20 09:23:19 crc kubenswrapper[4858]: I0320 09:23:19.315304 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnb9t" event={"ID":"0afbf51b-beb7-4808-8002-39a67f948cd1","Type":"ContainerStarted","Data":"e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce"} Mar 20 09:23:19 crc kubenswrapper[4858]: I0320 09:23:19.341957 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hnb9t" podStartSLOduration=3.351890069 podStartE2EDuration="7.341934957s" podCreationTimestamp="2026-03-20 09:23:12 +0000 UTC" firstStartedPulling="2026-03-20 09:23:14.260184516 +0000 UTC m=+1575.580602713" lastFinishedPulling="2026-03-20 09:23:18.250229414 +0000 UTC m=+1579.570647601" observedRunningTime="2026-03-20 09:23:19.337286962 +0000 UTC m=+1580.657705159" watchObservedRunningTime="2026-03-20 09:23:19.341934957 +0000 UTC m=+1580.662353154" Mar 20 09:23:22 crc kubenswrapper[4858]: I0320 09:23:22.768249 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:22 crc kubenswrapper[4858]: I0320 09:23:22.768680 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:23 crc kubenswrapper[4858]: I0320 09:23:23.810181 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hnb9t" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="registry-server" probeResult="failure" output=< Mar 20 09:23:23 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:23:23 crc kubenswrapper[4858]: > Mar 20 09:23:32 crc kubenswrapper[4858]: I0320 09:23:32.818103 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:32 crc kubenswrapper[4858]: I0320 09:23:32.867721 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:33 crc kubenswrapper[4858]: I0320 09:23:33.059643 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnb9t"] Mar 20 09:23:34 crc kubenswrapper[4858]: I0320 09:23:34.440381 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hnb9t" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="registry-server" containerID="cri-o://e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce" gracePeriod=2 Mar 20 09:23:34 crc kubenswrapper[4858]: I0320 09:23:34.873106 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:34 crc kubenswrapper[4858]: I0320 09:23:34.907216 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-utilities\") pod \"0afbf51b-beb7-4808-8002-39a67f948cd1\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " Mar 20 09:23:34 crc kubenswrapper[4858]: I0320 09:23:34.907725 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwtdc\" (UniqueName: \"kubernetes.io/projected/0afbf51b-beb7-4808-8002-39a67f948cd1-kube-api-access-fwtdc\") pod \"0afbf51b-beb7-4808-8002-39a67f948cd1\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " Mar 20 09:23:34 crc kubenswrapper[4858]: I0320 09:23:34.907824 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-catalog-content\") pod \"0afbf51b-beb7-4808-8002-39a67f948cd1\" (UID: \"0afbf51b-beb7-4808-8002-39a67f948cd1\") " Mar 20 09:23:34 crc kubenswrapper[4858]: I0320 09:23:34.911219 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-utilities" (OuterVolumeSpecName: "utilities") pod "0afbf51b-beb7-4808-8002-39a67f948cd1" (UID: "0afbf51b-beb7-4808-8002-39a67f948cd1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:23:34 crc kubenswrapper[4858]: I0320 09:23:34.918681 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0afbf51b-beb7-4808-8002-39a67f948cd1-kube-api-access-fwtdc" (OuterVolumeSpecName: "kube-api-access-fwtdc") pod "0afbf51b-beb7-4808-8002-39a67f948cd1" (UID: "0afbf51b-beb7-4808-8002-39a67f948cd1"). InnerVolumeSpecName "kube-api-access-fwtdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.009187 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.009270 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwtdc\" (UniqueName: \"kubernetes.io/projected/0afbf51b-beb7-4808-8002-39a67f948cd1-kube-api-access-fwtdc\") on node \"crc\" DevicePath \"\"" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.053459 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0afbf51b-beb7-4808-8002-39a67f948cd1" (UID: "0afbf51b-beb7-4808-8002-39a67f948cd1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.109813 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0afbf51b-beb7-4808-8002-39a67f948cd1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.450556 4858 generic.go:334] "Generic (PLEG): container finished" podID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerID="e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce" exitCode=0 Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.450631 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnb9t" event={"ID":"0afbf51b-beb7-4808-8002-39a67f948cd1","Type":"ContainerDied","Data":"e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce"} Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.450654 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnb9t" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.450694 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnb9t" event={"ID":"0afbf51b-beb7-4808-8002-39a67f948cd1","Type":"ContainerDied","Data":"7132aef89a55e7bb081c547cba02dbb98d2dd9b0ae7df9831f965b8465122d91"} Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.450725 4858 scope.go:117] "RemoveContainer" containerID="e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.477427 4858 scope.go:117] "RemoveContainer" containerID="23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.494226 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnb9t"] Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.499758 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hnb9t"] Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.508134 4858 scope.go:117] "RemoveContainer" containerID="225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.525573 4858 scope.go:117] "RemoveContainer" containerID="e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce" Mar 20 09:23:35 crc kubenswrapper[4858]: E0320 09:23:35.526010 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce\": container with ID starting with e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce not found: ID does not exist" containerID="e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.526046 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce"} err="failed to get container status \"e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce\": rpc error: code = NotFound desc = could not find container \"e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce\": container with ID starting with e68ba7c6397763a7ce4dfdd9a4b9c0392f4359789920f8ce14eedb5573db54ce not found: ID does not exist" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.526069 4858 scope.go:117] "RemoveContainer" containerID="23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd" Mar 20 09:23:35 crc kubenswrapper[4858]: E0320 09:23:35.526588 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd\": container with ID starting with 23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd not found: ID does not exist" containerID="23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.526622 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd"} err="failed to get container status \"23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd\": rpc error: code = NotFound desc = could not find container \"23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd\": container with ID starting with 23d196105c80e3ff0bf5fb337c85d7cfeebbe6e3acf7dd58b06faf20ed04e6fd not found: ID does not exist" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.526641 4858 scope.go:117] "RemoveContainer" containerID="225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f" Mar 20 09:23:35 crc kubenswrapper[4858]: E0320 09:23:35.526994 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f\": container with ID starting with 225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f not found: ID does not exist" containerID="225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f" Mar 20 09:23:35 crc kubenswrapper[4858]: I0320 09:23:35.527022 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f"} err="failed to get container status \"225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f\": rpc error: code = NotFound desc = could not find container \"225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f\": container with ID starting with 225a99362f3c51a01b8004048463867f703029036e24584831afcf92839d576f not found: ID does not exist" Mar 20 09:23:36 crc kubenswrapper[4858]: I0320 09:23:36.080642 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" path="/var/lib/kubelet/pods/0afbf51b-beb7-4808-8002-39a67f948cd1/volumes" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.144046 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566644-zktq5"] Mar 20 09:24:00 crc kubenswrapper[4858]: E0320 09:24:00.145413 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="extract-content" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.145435 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="extract-content" Mar 20 09:24:00 crc kubenswrapper[4858]: E0320 09:24:00.145453 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="extract-utilities" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.145461 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="extract-utilities" Mar 20 09:24:00 crc kubenswrapper[4858]: E0320 09:24:00.145489 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="registry-server" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.145500 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="registry-server" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.145701 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="0afbf51b-beb7-4808-8002-39a67f948cd1" containerName="registry-server" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.146635 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-zktq5" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.149500 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.149598 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.149499 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.151224 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-zktq5"] Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.195558 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52mcq\" (UniqueName: \"kubernetes.io/projected/beec987b-671a-477d-9386-ca8fd6099f86-kube-api-access-52mcq\") pod \"auto-csr-approver-29566644-zktq5\" (UID: \"beec987b-671a-477d-9386-ca8fd6099f86\") " pod="openshift-infra/auto-csr-approver-29566644-zktq5" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.297969 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52mcq\" (UniqueName: \"kubernetes.io/projected/beec987b-671a-477d-9386-ca8fd6099f86-kube-api-access-52mcq\") pod \"auto-csr-approver-29566644-zktq5\" (UID: \"beec987b-671a-477d-9386-ca8fd6099f86\") " pod="openshift-infra/auto-csr-approver-29566644-zktq5" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.319828 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52mcq\" (UniqueName: \"kubernetes.io/projected/beec987b-671a-477d-9386-ca8fd6099f86-kube-api-access-52mcq\") pod \"auto-csr-approver-29566644-zktq5\" (UID: \"beec987b-671a-477d-9386-ca8fd6099f86\") " pod="openshift-infra/auto-csr-approver-29566644-zktq5" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.477996 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-zktq5" Mar 20 09:24:00 crc kubenswrapper[4858]: I0320 09:24:00.902968 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-zktq5"] Mar 20 09:24:01 crc kubenswrapper[4858]: I0320 09:24:01.649559 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566644-zktq5" event={"ID":"beec987b-671a-477d-9386-ca8fd6099f86","Type":"ContainerStarted","Data":"adfd6234cf463c78bd38d05926a0d265a1a32bf23225f2a864f6885850d181a7"} Mar 20 09:24:02 crc kubenswrapper[4858]: I0320 09:24:02.657717 4858 generic.go:334] "Generic (PLEG): container finished" podID="beec987b-671a-477d-9386-ca8fd6099f86" containerID="9937a18668cc1e18839cbe1165390625fb2a949b666a28cb0c5f5637cae1712d" exitCode=0 Mar 20 09:24:02 crc kubenswrapper[4858]: I0320 09:24:02.657765 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566644-zktq5" event={"ID":"beec987b-671a-477d-9386-ca8fd6099f86","Type":"ContainerDied","Data":"9937a18668cc1e18839cbe1165390625fb2a949b666a28cb0c5f5637cae1712d"} Mar 20 09:24:03 crc kubenswrapper[4858]: I0320 09:24:03.970006 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-zktq5" Mar 20 09:24:04 crc kubenswrapper[4858]: I0320 09:24:04.053003 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52mcq\" (UniqueName: \"kubernetes.io/projected/beec987b-671a-477d-9386-ca8fd6099f86-kube-api-access-52mcq\") pod \"beec987b-671a-477d-9386-ca8fd6099f86\" (UID: \"beec987b-671a-477d-9386-ca8fd6099f86\") " Mar 20 09:24:04 crc kubenswrapper[4858]: I0320 09:24:04.064658 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beec987b-671a-477d-9386-ca8fd6099f86-kube-api-access-52mcq" (OuterVolumeSpecName: "kube-api-access-52mcq") pod "beec987b-671a-477d-9386-ca8fd6099f86" (UID: "beec987b-671a-477d-9386-ca8fd6099f86"). InnerVolumeSpecName "kube-api-access-52mcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:24:04 crc kubenswrapper[4858]: I0320 09:24:04.154920 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52mcq\" (UniqueName: \"kubernetes.io/projected/beec987b-671a-477d-9386-ca8fd6099f86-kube-api-access-52mcq\") on node \"crc\" DevicePath \"\"" Mar 20 09:24:04 crc kubenswrapper[4858]: I0320 09:24:04.673379 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566644-zktq5" event={"ID":"beec987b-671a-477d-9386-ca8fd6099f86","Type":"ContainerDied","Data":"adfd6234cf463c78bd38d05926a0d265a1a32bf23225f2a864f6885850d181a7"} Mar 20 09:24:04 crc kubenswrapper[4858]: I0320 09:24:04.673738 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adfd6234cf463c78bd38d05926a0d265a1a32bf23225f2a864f6885850d181a7" Mar 20 09:24:04 crc kubenswrapper[4858]: I0320 09:24:04.673417 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566644-zktq5" Mar 20 09:24:05 crc kubenswrapper[4858]: I0320 09:24:05.036304 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-nr9hv"] Mar 20 09:24:05 crc kubenswrapper[4858]: I0320 09:24:05.041674 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566638-nr9hv"] Mar 20 09:24:06 crc kubenswrapper[4858]: I0320 09:24:06.080231 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc" path="/var/lib/kubelet/pods/9901e88a-2409-4d2c-bfbc-2cdbdb76c1cc/volumes" Mar 20 09:24:30 crc kubenswrapper[4858]: I0320 09:24:30.372979 4858 scope.go:117] "RemoveContainer" containerID="8d4668e2a92f6c867a304e6f5397d6cf565dc1d8110e4e18a0772adad7914d90" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.257767 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-txhrn"] Mar 20 09:24:37 crc kubenswrapper[4858]: E0320 09:24:37.258482 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beec987b-671a-477d-9386-ca8fd6099f86" containerName="oc" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.258495 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="beec987b-671a-477d-9386-ca8fd6099f86" containerName="oc" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.258804 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="beec987b-671a-477d-9386-ca8fd6099f86" containerName="oc" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.260043 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.296003 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txhrn"] Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.423063 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-utilities\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.423292 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd7hr\" (UniqueName: \"kubernetes.io/projected/4382a2cd-3e16-448c-91c7-3299c9330ad4-kube-api-access-qd7hr\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.423465 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-catalog-content\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.524653 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-utilities\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.524750 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd7hr\" (UniqueName: \"kubernetes.io/projected/4382a2cd-3e16-448c-91c7-3299c9330ad4-kube-api-access-qd7hr\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.524811 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-catalog-content\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.525560 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-catalog-content\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.525769 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-utilities\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.757267 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd7hr\" (UniqueName: \"kubernetes.io/projected/4382a2cd-3e16-448c-91c7-3299c9330ad4-kube-api-access-qd7hr\") pod \"certified-operators-txhrn\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:37 crc kubenswrapper[4858]: I0320 09:24:37.889922 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:38 crc kubenswrapper[4858]: I0320 09:24:38.409456 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txhrn"] Mar 20 09:24:38 crc kubenswrapper[4858]: I0320 09:24:38.950335 4858 generic.go:334] "Generic (PLEG): container finished" podID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerID="bf22c2c078ed2099ef6f731a21bed6930897d8fafdbd5a53b687e41de8ae417c" exitCode=0 Mar 20 09:24:38 crc kubenswrapper[4858]: I0320 09:24:38.950379 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txhrn" event={"ID":"4382a2cd-3e16-448c-91c7-3299c9330ad4","Type":"ContainerDied","Data":"bf22c2c078ed2099ef6f731a21bed6930897d8fafdbd5a53b687e41de8ae417c"} Mar 20 09:24:38 crc kubenswrapper[4858]: I0320 09:24:38.950407 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txhrn" event={"ID":"4382a2cd-3e16-448c-91c7-3299c9330ad4","Type":"ContainerStarted","Data":"d659c8ca014f4ffed58db55396211bad41db32363315f6877a0af81b94f6f38d"} Mar 20 09:24:40 crc kubenswrapper[4858]: I0320 09:24:40.966345 4858 generic.go:334] "Generic (PLEG): container finished" podID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerID="7bf973d4f24cd6c39a20b260d80137ae3e5c22ce8853b15079636f93613857fe" exitCode=0 Mar 20 09:24:40 crc kubenswrapper[4858]: I0320 09:24:40.966392 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txhrn" event={"ID":"4382a2cd-3e16-448c-91c7-3299c9330ad4","Type":"ContainerDied","Data":"7bf973d4f24cd6c39a20b260d80137ae3e5c22ce8853b15079636f93613857fe"} Mar 20 09:24:41 crc kubenswrapper[4858]: I0320 09:24:41.984040 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txhrn" event={"ID":"4382a2cd-3e16-448c-91c7-3299c9330ad4","Type":"ContainerStarted","Data":"3beb811f79043d584b8abafcfa2e4ff595af6991bc5b61e4ea405287690d8a39"} Mar 20 09:24:42 crc kubenswrapper[4858]: I0320 09:24:42.015670 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-txhrn" podStartSLOduration=2.487374186 podStartE2EDuration="5.015649115s" podCreationTimestamp="2026-03-20 09:24:37 +0000 UTC" firstStartedPulling="2026-03-20 09:24:38.953298356 +0000 UTC m=+1660.273716553" lastFinishedPulling="2026-03-20 09:24:41.481573285 +0000 UTC m=+1662.801991482" observedRunningTime="2026-03-20 09:24:42.013400253 +0000 UTC m=+1663.333818460" watchObservedRunningTime="2026-03-20 09:24:42.015649115 +0000 UTC m=+1663.336067312" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.020229 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ng5jz"] Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.022341 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.029798 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ng5jz"] Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.184092 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmj7g\" (UniqueName: \"kubernetes.io/projected/0322fa4e-e861-4a7a-b7a3-ec60195a1292-kube-api-access-qmj7g\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.184677 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-utilities\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.184785 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-catalog-content\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.286212 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmj7g\" (UniqueName: \"kubernetes.io/projected/0322fa4e-e861-4a7a-b7a3-ec60195a1292-kube-api-access-qmj7g\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.286339 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-utilities\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.286369 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-catalog-content\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.287035 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-utilities\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.287087 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-catalog-content\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.307756 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmj7g\" (UniqueName: \"kubernetes.io/projected/0322fa4e-e861-4a7a-b7a3-ec60195a1292-kube-api-access-qmj7g\") pod \"community-operators-ng5jz\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.344058 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:46 crc kubenswrapper[4858]: I0320 09:24:46.862905 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ng5jz"] Mar 20 09:24:47 crc kubenswrapper[4858]: I0320 09:24:47.020494 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng5jz" event={"ID":"0322fa4e-e861-4a7a-b7a3-ec60195a1292","Type":"ContainerStarted","Data":"356fb4abef4c740821bd4708dad435f07babe867d2747ee4cdc8e019178191a4"} Mar 20 09:24:47 crc kubenswrapper[4858]: I0320 09:24:47.891044 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:47 crc kubenswrapper[4858]: I0320 09:24:47.891115 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:47 crc kubenswrapper[4858]: I0320 09:24:47.936013 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:48 crc kubenswrapper[4858]: I0320 09:24:48.028517 4858 generic.go:334] "Generic (PLEG): container finished" podID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerID="7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab" exitCode=0 Mar 20 09:24:48 crc kubenswrapper[4858]: I0320 09:24:48.028657 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng5jz" event={"ID":"0322fa4e-e861-4a7a-b7a3-ec60195a1292","Type":"ContainerDied","Data":"7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab"} Mar 20 09:24:48 crc kubenswrapper[4858]: I0320 09:24:48.081877 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:50 crc kubenswrapper[4858]: I0320 09:24:50.048997 4858 generic.go:334] "Generic (PLEG): container finished" podID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerID="ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5" exitCode=0 Mar 20 09:24:50 crc kubenswrapper[4858]: I0320 09:24:50.049154 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng5jz" event={"ID":"0322fa4e-e861-4a7a-b7a3-ec60195a1292","Type":"ContainerDied","Data":"ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5"} Mar 20 09:24:50 crc kubenswrapper[4858]: I0320 09:24:50.202559 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txhrn"] Mar 20 09:24:50 crc kubenswrapper[4858]: I0320 09:24:50.202803 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-txhrn" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="registry-server" containerID="cri-o://3beb811f79043d584b8abafcfa2e4ff595af6991bc5b61e4ea405287690d8a39" gracePeriod=2 Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.062506 4858 generic.go:334] "Generic (PLEG): container finished" podID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerID="3beb811f79043d584b8abafcfa2e4ff595af6991bc5b61e4ea405287690d8a39" exitCode=0 Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.062566 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txhrn" event={"ID":"4382a2cd-3e16-448c-91c7-3299c9330ad4","Type":"ContainerDied","Data":"3beb811f79043d584b8abafcfa2e4ff595af6991bc5b61e4ea405287690d8a39"} Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.703254 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.883098 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-utilities\") pod \"4382a2cd-3e16-448c-91c7-3299c9330ad4\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.883163 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-catalog-content\") pod \"4382a2cd-3e16-448c-91c7-3299c9330ad4\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.883189 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd7hr\" (UniqueName: \"kubernetes.io/projected/4382a2cd-3e16-448c-91c7-3299c9330ad4-kube-api-access-qd7hr\") pod \"4382a2cd-3e16-448c-91c7-3299c9330ad4\" (UID: \"4382a2cd-3e16-448c-91c7-3299c9330ad4\") " Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.885107 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-utilities" (OuterVolumeSpecName: "utilities") pod "4382a2cd-3e16-448c-91c7-3299c9330ad4" (UID: "4382a2cd-3e16-448c-91c7-3299c9330ad4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.888670 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4382a2cd-3e16-448c-91c7-3299c9330ad4-kube-api-access-qd7hr" (OuterVolumeSpecName: "kube-api-access-qd7hr") pod "4382a2cd-3e16-448c-91c7-3299c9330ad4" (UID: "4382a2cd-3e16-448c-91c7-3299c9330ad4"). InnerVolumeSpecName "kube-api-access-qd7hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.984390 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd7hr\" (UniqueName: \"kubernetes.io/projected/4382a2cd-3e16-448c-91c7-3299c9330ad4-kube-api-access-qd7hr\") on node \"crc\" DevicePath \"\"" Mar 20 09:24:51 crc kubenswrapper[4858]: I0320 09:24:51.984424 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.072964 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txhrn" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.082741 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txhrn" event={"ID":"4382a2cd-3e16-448c-91c7-3299c9330ad4","Type":"ContainerDied","Data":"d659c8ca014f4ffed58db55396211bad41db32363315f6877a0af81b94f6f38d"} Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.082799 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng5jz" event={"ID":"0322fa4e-e861-4a7a-b7a3-ec60195a1292","Type":"ContainerStarted","Data":"6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4"} Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.082836 4858 scope.go:117] "RemoveContainer" containerID="3beb811f79043d584b8abafcfa2e4ff595af6991bc5b61e4ea405287690d8a39" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.100819 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ng5jz" podStartSLOduration=2.9808794880000002 podStartE2EDuration="6.100768214s" podCreationTimestamp="2026-03-20 09:24:46 +0000 UTC" firstStartedPulling="2026-03-20 09:24:48.03038465 +0000 UTC m=+1669.350802857" lastFinishedPulling="2026-03-20 09:24:51.150273366 +0000 UTC m=+1672.470691583" observedRunningTime="2026-03-20 09:24:52.095428377 +0000 UTC m=+1673.415846584" watchObservedRunningTime="2026-03-20 09:24:52.100768214 +0000 UTC m=+1673.421186421" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.106840 4858 scope.go:117] "RemoveContainer" containerID="7bf973d4f24cd6c39a20b260d80137ae3e5c22ce8853b15079636f93613857fe" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.131808 4858 scope.go:117] "RemoveContainer" containerID="bf22c2c078ed2099ef6f731a21bed6930897d8fafdbd5a53b687e41de8ae417c" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.620355 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4382a2cd-3e16-448c-91c7-3299c9330ad4" (UID: "4382a2cd-3e16-448c-91c7-3299c9330ad4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.700125 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4382a2cd-3e16-448c-91c7-3299c9330ad4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.712492 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txhrn"] Mar 20 09:24:52 crc kubenswrapper[4858]: I0320 09:24:52.718944 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-txhrn"] Mar 20 09:24:54 crc kubenswrapper[4858]: I0320 09:24:54.079474 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" path="/var/lib/kubelet/pods/4382a2cd-3e16-448c-91c7-3299c9330ad4/volumes" Mar 20 09:24:56 crc kubenswrapper[4858]: I0320 09:24:56.344437 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:56 crc kubenswrapper[4858]: I0320 09:24:56.344840 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:56 crc kubenswrapper[4858]: I0320 09:24:56.390631 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:57 crc kubenswrapper[4858]: I0320 09:24:57.168856 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:57 crc kubenswrapper[4858]: I0320 09:24:57.598287 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ng5jz"] Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.141052 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ng5jz" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="registry-server" containerID="cri-o://6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4" gracePeriod=2 Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.553726 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.722607 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-catalog-content\") pod \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.722737 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmj7g\" (UniqueName: \"kubernetes.io/projected/0322fa4e-e861-4a7a-b7a3-ec60195a1292-kube-api-access-qmj7g\") pod \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.722784 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-utilities\") pod \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\" (UID: \"0322fa4e-e861-4a7a-b7a3-ec60195a1292\") " Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.723865 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-utilities" (OuterVolumeSpecName: "utilities") pod "0322fa4e-e861-4a7a-b7a3-ec60195a1292" (UID: "0322fa4e-e861-4a7a-b7a3-ec60195a1292"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.728783 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0322fa4e-e861-4a7a-b7a3-ec60195a1292-kube-api-access-qmj7g" (OuterVolumeSpecName: "kube-api-access-qmj7g") pod "0322fa4e-e861-4a7a-b7a3-ec60195a1292" (UID: "0322fa4e-e861-4a7a-b7a3-ec60195a1292"). InnerVolumeSpecName "kube-api-access-qmj7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.787139 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0322fa4e-e861-4a7a-b7a3-ec60195a1292" (UID: "0322fa4e-e861-4a7a-b7a3-ec60195a1292"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.825270 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.825338 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0322fa4e-e861-4a7a-b7a3-ec60195a1292-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:24:59 crc kubenswrapper[4858]: I0320 09:24:59.825358 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmj7g\" (UniqueName: \"kubernetes.io/projected/0322fa4e-e861-4a7a-b7a3-ec60195a1292-kube-api-access-qmj7g\") on node \"crc\" DevicePath \"\"" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.150867 4858 generic.go:334] "Generic (PLEG): container finished" podID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerID="6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4" exitCode=0 Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.150958 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ng5jz" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.150933 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng5jz" event={"ID":"0322fa4e-e861-4a7a-b7a3-ec60195a1292","Type":"ContainerDied","Data":"6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4"} Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.151757 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ng5jz" event={"ID":"0322fa4e-e861-4a7a-b7a3-ec60195a1292","Type":"ContainerDied","Data":"356fb4abef4c740821bd4708dad435f07babe867d2747ee4cdc8e019178191a4"} Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.151783 4858 scope.go:117] "RemoveContainer" containerID="6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.178516 4858 scope.go:117] "RemoveContainer" containerID="ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.180066 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ng5jz"] Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.185285 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ng5jz"] Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.200846 4858 scope.go:117] "RemoveContainer" containerID="7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.223913 4858 scope.go:117] "RemoveContainer" containerID="6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4" Mar 20 09:25:00 crc kubenswrapper[4858]: E0320 09:25:00.224357 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4\": container with ID starting with 6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4 not found: ID does not exist" containerID="6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.224394 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4"} err="failed to get container status \"6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4\": rpc error: code = NotFound desc = could not find container \"6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4\": container with ID starting with 6f9c2332c4305235665ad8d8ba5f4c1743185be99d3f3922066122db7d2a47d4 not found: ID does not exist" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.224416 4858 scope.go:117] "RemoveContainer" containerID="ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5" Mar 20 09:25:00 crc kubenswrapper[4858]: E0320 09:25:00.224806 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5\": container with ID starting with ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5 not found: ID does not exist" containerID="ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.224875 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5"} err="failed to get container status \"ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5\": rpc error: code = NotFound desc = could not find container \"ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5\": container with ID starting with ce70126066184428ff8af690de0e967e5af08dd6df081ed83d81ed1ff1b815c5 not found: ID does not exist" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.224928 4858 scope.go:117] "RemoveContainer" containerID="7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab" Mar 20 09:25:00 crc kubenswrapper[4858]: E0320 09:25:00.225602 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab\": container with ID starting with 7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab not found: ID does not exist" containerID="7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab" Mar 20 09:25:00 crc kubenswrapper[4858]: I0320 09:25:00.225704 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab"} err="failed to get container status \"7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab\": rpc error: code = NotFound desc = could not find container \"7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab\": container with ID starting with 7bf0cd4391b33e075931b3bf0cc51f630c9e9c009aa4574ffa69ceab6d8399ab not found: ID does not exist" Mar 20 09:25:02 crc kubenswrapper[4858]: I0320 09:25:02.079511 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" path="/var/lib/kubelet/pods/0322fa4e-e861-4a7a-b7a3-ec60195a1292/volumes" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.422065 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-srstb"] Mar 20 09:25:20 crc kubenswrapper[4858]: E0320 09:25:20.422805 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="registry-server" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.422818 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="registry-server" Mar 20 09:25:20 crc kubenswrapper[4858]: E0320 09:25:20.422834 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="extract-content" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.422840 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="extract-content" Mar 20 09:25:20 crc kubenswrapper[4858]: E0320 09:25:20.422851 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="registry-server" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.422858 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="registry-server" Mar 20 09:25:20 crc kubenswrapper[4858]: E0320 09:25:20.422869 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="extract-content" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.422875 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="extract-content" Mar 20 09:25:20 crc kubenswrapper[4858]: E0320 09:25:20.422904 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="extract-utilities" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.422910 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="extract-utilities" Mar 20 09:25:20 crc kubenswrapper[4858]: E0320 09:25:20.422920 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="extract-utilities" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.422928 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="extract-utilities" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.423096 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="0322fa4e-e861-4a7a-b7a3-ec60195a1292" containerName="registry-server" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.423112 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="4382a2cd-3e16-448c-91c7-3299c9330ad4" containerName="registry-server" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.424177 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.445220 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-srstb"] Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.497590 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgh2k\" (UniqueName: \"kubernetes.io/projected/a581e4d8-b717-4089-873a-5f5172c35dd4-kube-api-access-tgh2k\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.497668 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-utilities\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.497827 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-catalog-content\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.599120 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-catalog-content\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.599200 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgh2k\" (UniqueName: \"kubernetes.io/projected/a581e4d8-b717-4089-873a-5f5172c35dd4-kube-api-access-tgh2k\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.599258 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-utilities\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.599826 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-utilities\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.599958 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-catalog-content\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.618542 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgh2k\" (UniqueName: \"kubernetes.io/projected/a581e4d8-b717-4089-873a-5f5172c35dd4-kube-api-access-tgh2k\") pod \"redhat-marketplace-srstb\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:20 crc kubenswrapper[4858]: I0320 09:25:20.743745 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:21 crc kubenswrapper[4858]: I0320 09:25:21.182612 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-srstb"] Mar 20 09:25:21 crc kubenswrapper[4858]: I0320 09:25:21.311655 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srstb" event={"ID":"a581e4d8-b717-4089-873a-5f5172c35dd4","Type":"ContainerStarted","Data":"34920591aad4c9df7f8500efb6991c1b9028bb0e5cfc42947f4312c8eb0bbb1b"} Mar 20 09:25:22 crc kubenswrapper[4858]: I0320 09:25:22.322967 4858 generic.go:334] "Generic (PLEG): container finished" podID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerID="5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313" exitCode=0 Mar 20 09:25:22 crc kubenswrapper[4858]: I0320 09:25:22.323075 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srstb" event={"ID":"a581e4d8-b717-4089-873a-5f5172c35dd4","Type":"ContainerDied","Data":"5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313"} Mar 20 09:25:23 crc kubenswrapper[4858]: I0320 09:25:23.332349 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srstb" event={"ID":"a581e4d8-b717-4089-873a-5f5172c35dd4","Type":"ContainerStarted","Data":"8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06"} Mar 20 09:25:24 crc kubenswrapper[4858]: I0320 09:25:24.339825 4858 generic.go:334] "Generic (PLEG): container finished" podID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerID="8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06" exitCode=0 Mar 20 09:25:24 crc kubenswrapper[4858]: I0320 09:25:24.339895 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srstb" event={"ID":"a581e4d8-b717-4089-873a-5f5172c35dd4","Type":"ContainerDied","Data":"8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06"} Mar 20 09:25:25 crc kubenswrapper[4858]: I0320 09:25:25.349268 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srstb" event={"ID":"a581e4d8-b717-4089-873a-5f5172c35dd4","Type":"ContainerStarted","Data":"dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798"} Mar 20 09:25:25 crc kubenswrapper[4858]: I0320 09:25:25.366958 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-srstb" podStartSLOduration=2.932200167 podStartE2EDuration="5.366939037s" podCreationTimestamp="2026-03-20 09:25:20 +0000 UTC" firstStartedPulling="2026-03-20 09:25:22.326546153 +0000 UTC m=+1703.646964350" lastFinishedPulling="2026-03-20 09:25:24.761285023 +0000 UTC m=+1706.081703220" observedRunningTime="2026-03-20 09:25:25.364043248 +0000 UTC m=+1706.684461465" watchObservedRunningTime="2026-03-20 09:25:25.366939037 +0000 UTC m=+1706.687357234" Mar 20 09:25:30 crc kubenswrapper[4858]: I0320 09:25:30.744891 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:30 crc kubenswrapper[4858]: I0320 09:25:30.745847 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:30 crc kubenswrapper[4858]: I0320 09:25:30.790492 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:31 crc kubenswrapper[4858]: I0320 09:25:31.454426 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:31 crc kubenswrapper[4858]: I0320 09:25:31.503306 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-srstb"] Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.401589 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-srstb" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="registry-server" containerID="cri-o://dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798" gracePeriod=2 Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.807191 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.981205 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-catalog-content\") pod \"a581e4d8-b717-4089-873a-5f5172c35dd4\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.981268 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgh2k\" (UniqueName: \"kubernetes.io/projected/a581e4d8-b717-4089-873a-5f5172c35dd4-kube-api-access-tgh2k\") pod \"a581e4d8-b717-4089-873a-5f5172c35dd4\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.981450 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-utilities\") pod \"a581e4d8-b717-4089-873a-5f5172c35dd4\" (UID: \"a581e4d8-b717-4089-873a-5f5172c35dd4\") " Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.982135 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-utilities" (OuterVolumeSpecName: "utilities") pod "a581e4d8-b717-4089-873a-5f5172c35dd4" (UID: "a581e4d8-b717-4089-873a-5f5172c35dd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.982404 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:25:33 crc kubenswrapper[4858]: I0320 09:25:33.985706 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a581e4d8-b717-4089-873a-5f5172c35dd4-kube-api-access-tgh2k" (OuterVolumeSpecName: "kube-api-access-tgh2k") pod "a581e4d8-b717-4089-873a-5f5172c35dd4" (UID: "a581e4d8-b717-4089-873a-5f5172c35dd4"). InnerVolumeSpecName "kube-api-access-tgh2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.005912 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a581e4d8-b717-4089-873a-5f5172c35dd4" (UID: "a581e4d8-b717-4089-873a-5f5172c35dd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.083684 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a581e4d8-b717-4089-873a-5f5172c35dd4-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.083742 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgh2k\" (UniqueName: \"kubernetes.io/projected/a581e4d8-b717-4089-873a-5f5172c35dd4-kube-api-access-tgh2k\") on node \"crc\" DevicePath \"\"" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.413791 4858 generic.go:334] "Generic (PLEG): container finished" podID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerID="dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798" exitCode=0 Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.413833 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srstb" event={"ID":"a581e4d8-b717-4089-873a-5f5172c35dd4","Type":"ContainerDied","Data":"dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798"} Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.413853 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srstb" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.413875 4858 scope.go:117] "RemoveContainer" containerID="dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.413861 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srstb" event={"ID":"a581e4d8-b717-4089-873a-5f5172c35dd4","Type":"ContainerDied","Data":"34920591aad4c9df7f8500efb6991c1b9028bb0e5cfc42947f4312c8eb0bbb1b"} Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.444649 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-srstb"] Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.452248 4858 scope.go:117] "RemoveContainer" containerID="8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.452976 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-srstb"] Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.468072 4858 scope.go:117] "RemoveContainer" containerID="5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.501854 4858 scope.go:117] "RemoveContainer" containerID="dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798" Mar 20 09:25:34 crc kubenswrapper[4858]: E0320 09:25:34.502350 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798\": container with ID starting with dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798 not found: ID does not exist" containerID="dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.502405 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798"} err="failed to get container status \"dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798\": rpc error: code = NotFound desc = could not find container \"dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798\": container with ID starting with dc96c1bc902eb72ef2b480d8d44199ecb193be948abd22638c2d3d53bf52a798 not found: ID does not exist" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.502445 4858 scope.go:117] "RemoveContainer" containerID="8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06" Mar 20 09:25:34 crc kubenswrapper[4858]: E0320 09:25:34.503001 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06\": container with ID starting with 8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06 not found: ID does not exist" containerID="8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.503043 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06"} err="failed to get container status \"8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06\": rpc error: code = NotFound desc = could not find container \"8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06\": container with ID starting with 8816485b1c55d22779ddc419ed7a695f801a9d7ef47410d4e15100439f366a06 not found: ID does not exist" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.503075 4858 scope.go:117] "RemoveContainer" containerID="5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313" Mar 20 09:25:34 crc kubenswrapper[4858]: E0320 09:25:34.503356 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313\": container with ID starting with 5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313 not found: ID does not exist" containerID="5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313" Mar 20 09:25:34 crc kubenswrapper[4858]: I0320 09:25:34.503380 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313"} err="failed to get container status \"5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313\": rpc error: code = NotFound desc = could not find container \"5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313\": container with ID starting with 5ccec9c830ae73189d2897d77441a34234b9738f2777f5b5b9b9e161498fa313 not found: ID does not exist" Mar 20 09:25:36 crc kubenswrapper[4858]: I0320 09:25:36.079453 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" path="/var/lib/kubelet/pods/a581e4d8-b717-4089-873a-5f5172c35dd4/volumes" Mar 20 09:25:37 crc kubenswrapper[4858]: I0320 09:25:37.890800 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:25:37 crc kubenswrapper[4858]: I0320 09:25:37.891227 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.141529 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566646-bfnq6"] Mar 20 09:26:00 crc kubenswrapper[4858]: E0320 09:26:00.142302 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="extract-content" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.142331 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="extract-content" Mar 20 09:26:00 crc kubenswrapper[4858]: E0320 09:26:00.142344 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="registry-server" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.142350 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="registry-server" Mar 20 09:26:00 crc kubenswrapper[4858]: E0320 09:26:00.142377 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="extract-utilities" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.142383 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="extract-utilities" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.142511 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="a581e4d8-b717-4089-873a-5f5172c35dd4" containerName="registry-server" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.142950 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-bfnq6" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.146453 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.146654 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.146857 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.147032 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-bfnq6"] Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.198394 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fpzt\" (UniqueName: \"kubernetes.io/projected/b5e375c0-b15f-4f5a-8e38-ce13dec6f43d-kube-api-access-6fpzt\") pod \"auto-csr-approver-29566646-bfnq6\" (UID: \"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d\") " pod="openshift-infra/auto-csr-approver-29566646-bfnq6" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.299831 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fpzt\" (UniqueName: \"kubernetes.io/projected/b5e375c0-b15f-4f5a-8e38-ce13dec6f43d-kube-api-access-6fpzt\") pod \"auto-csr-approver-29566646-bfnq6\" (UID: \"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d\") " pod="openshift-infra/auto-csr-approver-29566646-bfnq6" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.323648 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fpzt\" (UniqueName: \"kubernetes.io/projected/b5e375c0-b15f-4f5a-8e38-ce13dec6f43d-kube-api-access-6fpzt\") pod \"auto-csr-approver-29566646-bfnq6\" (UID: \"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d\") " pod="openshift-infra/auto-csr-approver-29566646-bfnq6" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.514299 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-bfnq6" Mar 20 09:26:00 crc kubenswrapper[4858]: I0320 09:26:00.931164 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-bfnq6"] Mar 20 09:26:01 crc kubenswrapper[4858]: I0320 09:26:01.632483 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566646-bfnq6" event={"ID":"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d","Type":"ContainerStarted","Data":"06964d0dd5382f8deac05e784d57beee9674364e9d2d8f7410cbe9df86782c72"} Mar 20 09:26:02 crc kubenswrapper[4858]: I0320 09:26:02.640195 4858 generic.go:334] "Generic (PLEG): container finished" podID="b5e375c0-b15f-4f5a-8e38-ce13dec6f43d" containerID="aafae5f9549e3a8cb3d544f2625e181b97ee4243366ad88af1df7041860d6b82" exitCode=0 Mar 20 09:26:02 crc kubenswrapper[4858]: I0320 09:26:02.640256 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566646-bfnq6" event={"ID":"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d","Type":"ContainerDied","Data":"aafae5f9549e3a8cb3d544f2625e181b97ee4243366ad88af1df7041860d6b82"} Mar 20 09:26:03 crc kubenswrapper[4858]: I0320 09:26:03.963994 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-bfnq6" Mar 20 09:26:04 crc kubenswrapper[4858]: I0320 09:26:04.153040 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fpzt\" (UniqueName: \"kubernetes.io/projected/b5e375c0-b15f-4f5a-8e38-ce13dec6f43d-kube-api-access-6fpzt\") pod \"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d\" (UID: \"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d\") " Mar 20 09:26:04 crc kubenswrapper[4858]: I0320 09:26:04.158877 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5e375c0-b15f-4f5a-8e38-ce13dec6f43d-kube-api-access-6fpzt" (OuterVolumeSpecName: "kube-api-access-6fpzt") pod "b5e375c0-b15f-4f5a-8e38-ce13dec6f43d" (UID: "b5e375c0-b15f-4f5a-8e38-ce13dec6f43d"). InnerVolumeSpecName "kube-api-access-6fpzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:26:04 crc kubenswrapper[4858]: I0320 09:26:04.255028 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fpzt\" (UniqueName: \"kubernetes.io/projected/b5e375c0-b15f-4f5a-8e38-ce13dec6f43d-kube-api-access-6fpzt\") on node \"crc\" DevicePath \"\"" Mar 20 09:26:04 crc kubenswrapper[4858]: I0320 09:26:04.659780 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566646-bfnq6" event={"ID":"b5e375c0-b15f-4f5a-8e38-ce13dec6f43d","Type":"ContainerDied","Data":"06964d0dd5382f8deac05e784d57beee9674364e9d2d8f7410cbe9df86782c72"} Mar 20 09:26:04 crc kubenswrapper[4858]: I0320 09:26:04.659831 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566646-bfnq6" Mar 20 09:26:04 crc kubenswrapper[4858]: I0320 09:26:04.659846 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06964d0dd5382f8deac05e784d57beee9674364e9d2d8f7410cbe9df86782c72" Mar 20 09:26:05 crc kubenswrapper[4858]: I0320 09:26:05.033455 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-vpmsj"] Mar 20 09:26:05 crc kubenswrapper[4858]: I0320 09:26:05.039327 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566640-vpmsj"] Mar 20 09:26:06 crc kubenswrapper[4858]: I0320 09:26:06.088914 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba8e3af1-a5eb-4855-a9c1-903713dac8c2" path="/var/lib/kubelet/pods/ba8e3af1-a5eb-4855-a9c1-903713dac8c2/volumes" Mar 20 09:26:07 crc kubenswrapper[4858]: I0320 09:26:07.890221 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:26:07 crc kubenswrapper[4858]: I0320 09:26:07.890303 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:26:30 crc kubenswrapper[4858]: I0320 09:26:30.513476 4858 scope.go:117] "RemoveContainer" containerID="4a94e79ae5b566aacad8414fae2ea0b962213b680ca0078c4a63944122d88e95" Mar 20 09:26:37 crc kubenswrapper[4858]: I0320 09:26:37.890963 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:26:37 crc kubenswrapper[4858]: I0320 09:26:37.891649 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:26:37 crc kubenswrapper[4858]: I0320 09:26:37.891711 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:26:37 crc kubenswrapper[4858]: I0320 09:26:37.892604 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:26:37 crc kubenswrapper[4858]: I0320 09:26:37.892670 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" gracePeriod=600 Mar 20 09:26:38 crc kubenswrapper[4858]: E0320 09:26:38.068961 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:26:38 crc kubenswrapper[4858]: I0320 09:26:38.950175 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" exitCode=0 Mar 20 09:26:38 crc kubenswrapper[4858]: I0320 09:26:38.950243 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea"} Mar 20 09:26:38 crc kubenswrapper[4858]: I0320 09:26:38.950299 4858 scope.go:117] "RemoveContainer" containerID="1214de77b7ba6f5035904ddcf50d5dc1a7d89457797a1d34a3dfae1bc23c2fd0" Mar 20 09:26:38 crc kubenswrapper[4858]: I0320 09:26:38.951096 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:26:38 crc kubenswrapper[4858]: E0320 09:26:38.951600 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:26:50 crc kubenswrapper[4858]: I0320 09:26:50.075442 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:26:50 crc kubenswrapper[4858]: E0320 09:26:50.076351 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:27:02 crc kubenswrapper[4858]: I0320 09:27:02.070599 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:27:02 crc kubenswrapper[4858]: E0320 09:27:02.071602 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:27:13 crc kubenswrapper[4858]: I0320 09:27:13.070727 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:27:13 crc kubenswrapper[4858]: E0320 09:27:13.071658 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:27:27 crc kubenswrapper[4858]: I0320 09:27:27.071059 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:27:27 crc kubenswrapper[4858]: E0320 09:27:27.072269 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:27:40 crc kubenswrapper[4858]: I0320 09:27:40.074229 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:27:40 crc kubenswrapper[4858]: E0320 09:27:40.075093 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:27:54 crc kubenswrapper[4858]: I0320 09:27:54.070691 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:27:54 crc kubenswrapper[4858]: E0320 09:27:54.072844 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.144430 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566648-dc294"] Mar 20 09:28:00 crc kubenswrapper[4858]: E0320 09:28:00.145175 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5e375c0-b15f-4f5a-8e38-ce13dec6f43d" containerName="oc" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.145191 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5e375c0-b15f-4f5a-8e38-ce13dec6f43d" containerName="oc" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.145517 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5e375c0-b15f-4f5a-8e38-ce13dec6f43d" containerName="oc" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.146067 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-dc294" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.148428 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.149284 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.149352 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.150528 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-dc294"] Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.214096 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvdxd\" (UniqueName: \"kubernetes.io/projected/f063b8c5-82cd-4d57-96f5-b91adc5717b2-kube-api-access-xvdxd\") pod \"auto-csr-approver-29566648-dc294\" (UID: \"f063b8c5-82cd-4d57-96f5-b91adc5717b2\") " pod="openshift-infra/auto-csr-approver-29566648-dc294" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.315911 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvdxd\" (UniqueName: \"kubernetes.io/projected/f063b8c5-82cd-4d57-96f5-b91adc5717b2-kube-api-access-xvdxd\") pod \"auto-csr-approver-29566648-dc294\" (UID: \"f063b8c5-82cd-4d57-96f5-b91adc5717b2\") " pod="openshift-infra/auto-csr-approver-29566648-dc294" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.341536 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvdxd\" (UniqueName: \"kubernetes.io/projected/f063b8c5-82cd-4d57-96f5-b91adc5717b2-kube-api-access-xvdxd\") pod \"auto-csr-approver-29566648-dc294\" (UID: \"f063b8c5-82cd-4d57-96f5-b91adc5717b2\") " pod="openshift-infra/auto-csr-approver-29566648-dc294" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.466561 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-dc294" Mar 20 09:28:00 crc kubenswrapper[4858]: I0320 09:28:00.892628 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-dc294"] Mar 20 09:28:01 crc kubenswrapper[4858]: I0320 09:28:01.615830 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566648-dc294" event={"ID":"f063b8c5-82cd-4d57-96f5-b91adc5717b2","Type":"ContainerStarted","Data":"5ba60b0d0d4a701b9ba8ec8dcfa473d74645985f49d9455f3f392b63bdc5cf49"} Mar 20 09:28:02 crc kubenswrapper[4858]: I0320 09:28:02.626505 4858 generic.go:334] "Generic (PLEG): container finished" podID="f063b8c5-82cd-4d57-96f5-b91adc5717b2" containerID="686926f47c91f575054458d02df1066f27f634331ffea7a6d74facb2a5c63e46" exitCode=0 Mar 20 09:28:02 crc kubenswrapper[4858]: I0320 09:28:02.626934 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566648-dc294" event={"ID":"f063b8c5-82cd-4d57-96f5-b91adc5717b2","Type":"ContainerDied","Data":"686926f47c91f575054458d02df1066f27f634331ffea7a6d74facb2a5c63e46"} Mar 20 09:28:03 crc kubenswrapper[4858]: I0320 09:28:03.951425 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-dc294" Mar 20 09:28:04 crc kubenswrapper[4858]: I0320 09:28:04.006112 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvdxd\" (UniqueName: \"kubernetes.io/projected/f063b8c5-82cd-4d57-96f5-b91adc5717b2-kube-api-access-xvdxd\") pod \"f063b8c5-82cd-4d57-96f5-b91adc5717b2\" (UID: \"f063b8c5-82cd-4d57-96f5-b91adc5717b2\") " Mar 20 09:28:04 crc kubenswrapper[4858]: I0320 09:28:04.015535 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f063b8c5-82cd-4d57-96f5-b91adc5717b2-kube-api-access-xvdxd" (OuterVolumeSpecName: "kube-api-access-xvdxd") pod "f063b8c5-82cd-4d57-96f5-b91adc5717b2" (UID: "f063b8c5-82cd-4d57-96f5-b91adc5717b2"). InnerVolumeSpecName "kube-api-access-xvdxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:28:04 crc kubenswrapper[4858]: I0320 09:28:04.108003 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvdxd\" (UniqueName: \"kubernetes.io/projected/f063b8c5-82cd-4d57-96f5-b91adc5717b2-kube-api-access-xvdxd\") on node \"crc\" DevicePath \"\"" Mar 20 09:28:04 crc kubenswrapper[4858]: I0320 09:28:04.644873 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566648-dc294" event={"ID":"f063b8c5-82cd-4d57-96f5-b91adc5717b2","Type":"ContainerDied","Data":"5ba60b0d0d4a701b9ba8ec8dcfa473d74645985f49d9455f3f392b63bdc5cf49"} Mar 20 09:28:04 crc kubenswrapper[4858]: I0320 09:28:04.644923 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ba60b0d0d4a701b9ba8ec8dcfa473d74645985f49d9455f3f392b63bdc5cf49" Mar 20 09:28:04 crc kubenswrapper[4858]: I0320 09:28:04.644970 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566648-dc294" Mar 20 09:28:05 crc kubenswrapper[4858]: I0320 09:28:05.023744 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-rlsxc"] Mar 20 09:28:05 crc kubenswrapper[4858]: I0320 09:28:05.029473 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566642-rlsxc"] Mar 20 09:28:06 crc kubenswrapper[4858]: I0320 09:28:06.070645 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:28:06 crc kubenswrapper[4858]: E0320 09:28:06.070905 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:28:06 crc kubenswrapper[4858]: I0320 09:28:06.081110 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adae7626-9eb0-46d9-b8d8-d5d0629181b4" path="/var/lib/kubelet/pods/adae7626-9eb0-46d9-b8d8-d5d0629181b4/volumes" Mar 20 09:28:20 crc kubenswrapper[4858]: I0320 09:28:20.074157 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:28:20 crc kubenswrapper[4858]: E0320 09:28:20.075204 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:28:30 crc kubenswrapper[4858]: I0320 09:28:30.601025 4858 scope.go:117] "RemoveContainer" containerID="c76c5b41a4e11da041a4d4407fbad14b2e0953d23cc0c9a95d6449fe175bf58b" Mar 20 09:28:34 crc kubenswrapper[4858]: I0320 09:28:34.070540 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:28:34 crc kubenswrapper[4858]: E0320 09:28:34.071734 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:28:48 crc kubenswrapper[4858]: I0320 09:28:48.070028 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:28:48 crc kubenswrapper[4858]: E0320 09:28:48.072122 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:29:02 crc kubenswrapper[4858]: I0320 09:29:02.071425 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:29:02 crc kubenswrapper[4858]: E0320 09:29:02.072448 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:29:14 crc kubenswrapper[4858]: I0320 09:29:14.070032 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:29:14 crc kubenswrapper[4858]: E0320 09:29:14.070931 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:29:26 crc kubenswrapper[4858]: I0320 09:29:26.070751 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:29:26 crc kubenswrapper[4858]: E0320 09:29:26.072239 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:29:38 crc kubenswrapper[4858]: I0320 09:29:38.070461 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:29:38 crc kubenswrapper[4858]: E0320 09:29:38.071583 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:29:50 crc kubenswrapper[4858]: I0320 09:29:50.075282 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:29:50 crc kubenswrapper[4858]: E0320 09:29:50.076599 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.207263 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566650-prn7d"] Mar 20 09:30:00 crc kubenswrapper[4858]: E0320 09:30:00.213784 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f063b8c5-82cd-4d57-96f5-b91adc5717b2" containerName="oc" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.213826 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="f063b8c5-82cd-4d57-96f5-b91adc5717b2" containerName="oc" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.214074 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="f063b8c5-82cd-4d57-96f5-b91adc5717b2" containerName="oc" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.214883 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-prn7d" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.216846 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq"] Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.218518 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.226181 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.226280 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.226859 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.227606 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.227768 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-prn7d"] Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.234022 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.234311 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq"] Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.341283 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szk9x\" (UniqueName: \"kubernetes.io/projected/c61c56a2-c37b-4b49-b970-0da0a98defcc-kube-api-access-szk9x\") pod \"auto-csr-approver-29566650-prn7d\" (UID: \"c61c56a2-c37b-4b49-b970-0da0a98defcc\") " pod="openshift-infra/auto-csr-approver-29566650-prn7d" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.341360 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpl2m\" (UniqueName: \"kubernetes.io/projected/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-kube-api-access-jpl2m\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.341400 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-config-volume\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.341480 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-secret-volume\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.443483 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szk9x\" (UniqueName: \"kubernetes.io/projected/c61c56a2-c37b-4b49-b970-0da0a98defcc-kube-api-access-szk9x\") pod \"auto-csr-approver-29566650-prn7d\" (UID: \"c61c56a2-c37b-4b49-b970-0da0a98defcc\") " pod="openshift-infra/auto-csr-approver-29566650-prn7d" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.443543 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpl2m\" (UniqueName: \"kubernetes.io/projected/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-kube-api-access-jpl2m\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.443579 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-config-volume\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.443629 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-secret-volume\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.445849 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-config-volume\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.464515 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-secret-volume\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.466408 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szk9x\" (UniqueName: \"kubernetes.io/projected/c61c56a2-c37b-4b49-b970-0da0a98defcc-kube-api-access-szk9x\") pod \"auto-csr-approver-29566650-prn7d\" (UID: \"c61c56a2-c37b-4b49-b970-0da0a98defcc\") " pod="openshift-infra/auto-csr-approver-29566650-prn7d" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.467204 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpl2m\" (UniqueName: \"kubernetes.io/projected/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-kube-api-access-jpl2m\") pod \"collect-profiles-29566650-2w9wq\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.577377 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-prn7d" Mar 20 09:30:00 crc kubenswrapper[4858]: I0320 09:30:00.590300 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.057301 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-prn7d"] Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.068299 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.070243 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:30:01 crc kubenswrapper[4858]: E0320 09:30:01.070548 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:30:01 crc kubenswrapper[4858]: W0320 09:30:01.129360 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b668fba_f2fa_4f78_8cd4_5f840b4b0c68.slice/crio-e2d0000c34e4e3f5600987e0ff36ee99b517c5cdbca09b760dbba4a9389687c0 WatchSource:0}: Error finding container e2d0000c34e4e3f5600987e0ff36ee99b517c5cdbca09b760dbba4a9389687c0: Status 404 returned error can't find the container with id e2d0000c34e4e3f5600987e0ff36ee99b517c5cdbca09b760dbba4a9389687c0 Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.131058 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq"] Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.594026 4858 generic.go:334] "Generic (PLEG): container finished" podID="6b668fba-f2fa-4f78-8cd4-5f840b4b0c68" containerID="3da8290bf94a10fd912babe6eebd83b1e430b382ceb7685d1d1c6949305e3222" exitCode=0 Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.594109 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" event={"ID":"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68","Type":"ContainerDied","Data":"3da8290bf94a10fd912babe6eebd83b1e430b382ceb7685d1d1c6949305e3222"} Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.594146 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" event={"ID":"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68","Type":"ContainerStarted","Data":"e2d0000c34e4e3f5600987e0ff36ee99b517c5cdbca09b760dbba4a9389687c0"} Mar 20 09:30:01 crc kubenswrapper[4858]: I0320 09:30:01.595616 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566650-prn7d" event={"ID":"c61c56a2-c37b-4b49-b970-0da0a98defcc","Type":"ContainerStarted","Data":"525ca574f4d45b9e74ecc9939e416bac5735dddd1102c763d1f109fe377b3595"} Mar 20 09:30:02 crc kubenswrapper[4858]: I0320 09:30:02.913390 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:02 crc kubenswrapper[4858]: I0320 09:30:02.996127 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpl2m\" (UniqueName: \"kubernetes.io/projected/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-kube-api-access-jpl2m\") pod \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " Mar 20 09:30:02 crc kubenswrapper[4858]: I0320 09:30:02.996265 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-secret-volume\") pod \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " Mar 20 09:30:02 crc kubenswrapper[4858]: I0320 09:30:02.996418 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-config-volume\") pod \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\" (UID: \"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68\") " Mar 20 09:30:02 crc kubenswrapper[4858]: I0320 09:30:02.997578 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-config-volume" (OuterVolumeSpecName: "config-volume") pod "6b668fba-f2fa-4f78-8cd4-5f840b4b0c68" (UID: "6b668fba-f2fa-4f78-8cd4-5f840b4b0c68"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.003419 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-kube-api-access-jpl2m" (OuterVolumeSpecName: "kube-api-access-jpl2m") pod "6b668fba-f2fa-4f78-8cd4-5f840b4b0c68" (UID: "6b668fba-f2fa-4f78-8cd4-5f840b4b0c68"). InnerVolumeSpecName "kube-api-access-jpl2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.003548 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6b668fba-f2fa-4f78-8cd4-5f840b4b0c68" (UID: "6b668fba-f2fa-4f78-8cd4-5f840b4b0c68"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.099702 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpl2m\" (UniqueName: \"kubernetes.io/projected/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-kube-api-access-jpl2m\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.100300 4858 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.100339 4858 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6b668fba-f2fa-4f78-8cd4-5f840b4b0c68-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.613249 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" event={"ID":"6b668fba-f2fa-4f78-8cd4-5f840b4b0c68","Type":"ContainerDied","Data":"e2d0000c34e4e3f5600987e0ff36ee99b517c5cdbca09b760dbba4a9389687c0"} Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.613305 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2d0000c34e4e3f5600987e0ff36ee99b517c5cdbca09b760dbba4a9389687c0" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.613411 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566650-2w9wq" Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.616060 4858 generic.go:334] "Generic (PLEG): container finished" podID="c61c56a2-c37b-4b49-b970-0da0a98defcc" containerID="4f40169bac1ee45370466ab61edba74c94ad2f30f2ed5563f57e2a6a2fa2c48b" exitCode=0 Mar 20 09:30:03 crc kubenswrapper[4858]: I0320 09:30:03.616116 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566650-prn7d" event={"ID":"c61c56a2-c37b-4b49-b970-0da0a98defcc","Type":"ContainerDied","Data":"4f40169bac1ee45370466ab61edba74c94ad2f30f2ed5563f57e2a6a2fa2c48b"} Mar 20 09:30:04 crc kubenswrapper[4858]: I0320 09:30:04.924918 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-prn7d" Mar 20 09:30:05 crc kubenswrapper[4858]: I0320 09:30:05.029956 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szk9x\" (UniqueName: \"kubernetes.io/projected/c61c56a2-c37b-4b49-b970-0da0a98defcc-kube-api-access-szk9x\") pod \"c61c56a2-c37b-4b49-b970-0da0a98defcc\" (UID: \"c61c56a2-c37b-4b49-b970-0da0a98defcc\") " Mar 20 09:30:05 crc kubenswrapper[4858]: I0320 09:30:05.037381 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61c56a2-c37b-4b49-b970-0da0a98defcc-kube-api-access-szk9x" (OuterVolumeSpecName: "kube-api-access-szk9x") pod "c61c56a2-c37b-4b49-b970-0da0a98defcc" (UID: "c61c56a2-c37b-4b49-b970-0da0a98defcc"). InnerVolumeSpecName "kube-api-access-szk9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:30:05 crc kubenswrapper[4858]: I0320 09:30:05.132255 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szk9x\" (UniqueName: \"kubernetes.io/projected/c61c56a2-c37b-4b49-b970-0da0a98defcc-kube-api-access-szk9x\") on node \"crc\" DevicePath \"\"" Mar 20 09:30:05 crc kubenswrapper[4858]: I0320 09:30:05.634136 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566650-prn7d" event={"ID":"c61c56a2-c37b-4b49-b970-0da0a98defcc","Type":"ContainerDied","Data":"525ca574f4d45b9e74ecc9939e416bac5735dddd1102c763d1f109fe377b3595"} Mar 20 09:30:05 crc kubenswrapper[4858]: I0320 09:30:05.634198 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566650-prn7d" Mar 20 09:30:05 crc kubenswrapper[4858]: I0320 09:30:05.634200 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="525ca574f4d45b9e74ecc9939e416bac5735dddd1102c763d1f109fe377b3595" Mar 20 09:30:06 crc kubenswrapper[4858]: I0320 09:30:06.009052 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-zktq5"] Mar 20 09:30:06 crc kubenswrapper[4858]: I0320 09:30:06.016255 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566644-zktq5"] Mar 20 09:30:06 crc kubenswrapper[4858]: I0320 09:30:06.084529 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beec987b-671a-477d-9386-ca8fd6099f86" path="/var/lib/kubelet/pods/beec987b-671a-477d-9386-ca8fd6099f86/volumes" Mar 20 09:30:12 crc kubenswrapper[4858]: I0320 09:30:12.070674 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:30:12 crc kubenswrapper[4858]: E0320 09:30:12.071867 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:30:23 crc kubenswrapper[4858]: I0320 09:30:23.071029 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:30:23 crc kubenswrapper[4858]: E0320 09:30:23.072012 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:30:30 crc kubenswrapper[4858]: I0320 09:30:30.717138 4858 scope.go:117] "RemoveContainer" containerID="9937a18668cc1e18839cbe1165390625fb2a949b666a28cb0c5f5637cae1712d" Mar 20 09:30:36 crc kubenswrapper[4858]: I0320 09:30:36.071066 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:30:36 crc kubenswrapper[4858]: E0320 09:30:36.072004 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:30:49 crc kubenswrapper[4858]: I0320 09:30:49.070589 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:30:49 crc kubenswrapper[4858]: E0320 09:30:49.071766 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:31:02 crc kubenswrapper[4858]: I0320 09:31:02.070537 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:31:02 crc kubenswrapper[4858]: E0320 09:31:02.072434 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:31:14 crc kubenswrapper[4858]: I0320 09:31:14.070430 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:31:14 crc kubenswrapper[4858]: E0320 09:31:14.072657 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:31:25 crc kubenswrapper[4858]: I0320 09:31:25.071002 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:31:25 crc kubenswrapper[4858]: E0320 09:31:25.072197 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:31:36 crc kubenswrapper[4858]: I0320 09:31:36.070342 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:31:36 crc kubenswrapper[4858]: E0320 09:31:36.071116 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:31:47 crc kubenswrapper[4858]: I0320 09:31:47.070718 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:31:47 crc kubenswrapper[4858]: I0320 09:31:47.541214 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"5817ed13dcb1fb797014d77cda07222984a083879566a3f0d5db8e9e973bd960"} Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.154726 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566652-nzq49"] Mar 20 09:32:00 crc kubenswrapper[4858]: E0320 09:32:00.155893 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61c56a2-c37b-4b49-b970-0da0a98defcc" containerName="oc" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.155914 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61c56a2-c37b-4b49-b970-0da0a98defcc" containerName="oc" Mar 20 09:32:00 crc kubenswrapper[4858]: E0320 09:32:00.155929 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b668fba-f2fa-4f78-8cd4-5f840b4b0c68" containerName="collect-profiles" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.155936 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b668fba-f2fa-4f78-8cd4-5f840b4b0c68" containerName="collect-profiles" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.156080 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61c56a2-c37b-4b49-b970-0da0a98defcc" containerName="oc" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.156100 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b668fba-f2fa-4f78-8cd4-5f840b4b0c68" containerName="collect-profiles" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.156738 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-nzq49" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.160622 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.160640 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.160949 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.172402 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-nzq49"] Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.185429 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh2rb\" (UniqueName: \"kubernetes.io/projected/773714cb-9045-4504-8813-9f389ed64883-kube-api-access-gh2rb\") pod \"auto-csr-approver-29566652-nzq49\" (UID: \"773714cb-9045-4504-8813-9f389ed64883\") " pod="openshift-infra/auto-csr-approver-29566652-nzq49" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.286140 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh2rb\" (UniqueName: \"kubernetes.io/projected/773714cb-9045-4504-8813-9f389ed64883-kube-api-access-gh2rb\") pod \"auto-csr-approver-29566652-nzq49\" (UID: \"773714cb-9045-4504-8813-9f389ed64883\") " pod="openshift-infra/auto-csr-approver-29566652-nzq49" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.312622 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh2rb\" (UniqueName: \"kubernetes.io/projected/773714cb-9045-4504-8813-9f389ed64883-kube-api-access-gh2rb\") pod \"auto-csr-approver-29566652-nzq49\" (UID: \"773714cb-9045-4504-8813-9f389ed64883\") " pod="openshift-infra/auto-csr-approver-29566652-nzq49" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.483788 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-nzq49" Mar 20 09:32:00 crc kubenswrapper[4858]: I0320 09:32:00.741786 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-nzq49"] Mar 20 09:32:01 crc kubenswrapper[4858]: I0320 09:32:01.659893 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566652-nzq49" event={"ID":"773714cb-9045-4504-8813-9f389ed64883","Type":"ContainerStarted","Data":"ab9245e3ad5522655c9689f094d1a3a88bbce5083001cef802c698d696a7c4da"} Mar 20 09:32:02 crc kubenswrapper[4858]: I0320 09:32:02.670722 4858 generic.go:334] "Generic (PLEG): container finished" podID="773714cb-9045-4504-8813-9f389ed64883" containerID="2388726181deac0f3720191eea32ad01bbcb2388a8f20ae64a1684ce7835bd53" exitCode=0 Mar 20 09:32:02 crc kubenswrapper[4858]: I0320 09:32:02.670959 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566652-nzq49" event={"ID":"773714cb-9045-4504-8813-9f389ed64883","Type":"ContainerDied","Data":"2388726181deac0f3720191eea32ad01bbcb2388a8f20ae64a1684ce7835bd53"} Mar 20 09:32:03 crc kubenswrapper[4858]: I0320 09:32:03.995294 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-nzq49" Mar 20 09:32:04 crc kubenswrapper[4858]: I0320 09:32:04.152854 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh2rb\" (UniqueName: \"kubernetes.io/projected/773714cb-9045-4504-8813-9f389ed64883-kube-api-access-gh2rb\") pod \"773714cb-9045-4504-8813-9f389ed64883\" (UID: \"773714cb-9045-4504-8813-9f389ed64883\") " Mar 20 09:32:04 crc kubenswrapper[4858]: I0320 09:32:04.161525 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773714cb-9045-4504-8813-9f389ed64883-kube-api-access-gh2rb" (OuterVolumeSpecName: "kube-api-access-gh2rb") pod "773714cb-9045-4504-8813-9f389ed64883" (UID: "773714cb-9045-4504-8813-9f389ed64883"). InnerVolumeSpecName "kube-api-access-gh2rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:32:04 crc kubenswrapper[4858]: I0320 09:32:04.255333 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh2rb\" (UniqueName: \"kubernetes.io/projected/773714cb-9045-4504-8813-9f389ed64883-kube-api-access-gh2rb\") on node \"crc\" DevicePath \"\"" Mar 20 09:32:04 crc kubenswrapper[4858]: I0320 09:32:04.692534 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566652-nzq49" event={"ID":"773714cb-9045-4504-8813-9f389ed64883","Type":"ContainerDied","Data":"ab9245e3ad5522655c9689f094d1a3a88bbce5083001cef802c698d696a7c4da"} Mar 20 09:32:04 crc kubenswrapper[4858]: I0320 09:32:04.693078 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab9245e3ad5522655c9689f094d1a3a88bbce5083001cef802c698d696a7c4da" Mar 20 09:32:04 crc kubenswrapper[4858]: I0320 09:32:04.692589 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566652-nzq49" Mar 20 09:32:05 crc kubenswrapper[4858]: I0320 09:32:05.074239 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-bfnq6"] Mar 20 09:32:05 crc kubenswrapper[4858]: I0320 09:32:05.081257 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566646-bfnq6"] Mar 20 09:32:06 crc kubenswrapper[4858]: I0320 09:32:06.082217 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5e375c0-b15f-4f5a-8e38-ce13dec6f43d" path="/var/lib/kubelet/pods/b5e375c0-b15f-4f5a-8e38-ce13dec6f43d/volumes" Mar 20 09:32:30 crc kubenswrapper[4858]: I0320 09:32:30.811509 4858 scope.go:117] "RemoveContainer" containerID="aafae5f9549e3a8cb3d544f2625e181b97ee4243366ad88af1df7041860d6b82" Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.863450 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8hv5z"] Mar 20 09:33:17 crc kubenswrapper[4858]: E0320 09:33:17.864211 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773714cb-9045-4504-8813-9f389ed64883" containerName="oc" Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.864227 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="773714cb-9045-4504-8813-9f389ed64883" containerName="oc" Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.864546 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="773714cb-9045-4504-8813-9f389ed64883" containerName="oc" Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.865779 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.875604 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8hv5z"] Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.905907 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-utilities\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.906004 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfrzg\" (UniqueName: \"kubernetes.io/projected/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-kube-api-access-wfrzg\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:17 crc kubenswrapper[4858]: I0320 09:33:17.906045 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-catalog-content\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.007317 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-utilities\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.007811 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfrzg\" (UniqueName: \"kubernetes.io/projected/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-kube-api-access-wfrzg\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.007915 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-catalog-content\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.007947 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-utilities\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.008179 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-catalog-content\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.031766 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfrzg\" (UniqueName: \"kubernetes.io/projected/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-kube-api-access-wfrzg\") pod \"redhat-operators-8hv5z\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.190521 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:18 crc kubenswrapper[4858]: I0320 09:33:18.704114 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8hv5z"] Mar 20 09:33:19 crc kubenswrapper[4858]: I0320 09:33:19.298502 4858 generic.go:334] "Generic (PLEG): container finished" podID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerID="ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f" exitCode=0 Mar 20 09:33:19 crc kubenswrapper[4858]: I0320 09:33:19.298611 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hv5z" event={"ID":"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f","Type":"ContainerDied","Data":"ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f"} Mar 20 09:33:19 crc kubenswrapper[4858]: I0320 09:33:19.299073 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hv5z" event={"ID":"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f","Type":"ContainerStarted","Data":"665ae1ffccb41796828550a6c79fff244d02c5573e2ce8648bcd4e06ab04f97e"} Mar 20 09:33:20 crc kubenswrapper[4858]: I0320 09:33:20.309325 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hv5z" event={"ID":"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f","Type":"ContainerStarted","Data":"78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe"} Mar 20 09:33:21 crc kubenswrapper[4858]: I0320 09:33:21.319205 4858 generic.go:334] "Generic (PLEG): container finished" podID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerID="78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe" exitCode=0 Mar 20 09:33:21 crc kubenswrapper[4858]: I0320 09:33:21.319284 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hv5z" event={"ID":"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f","Type":"ContainerDied","Data":"78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe"} Mar 20 09:33:22 crc kubenswrapper[4858]: I0320 09:33:22.333955 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hv5z" event={"ID":"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f","Type":"ContainerStarted","Data":"77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1"} Mar 20 09:33:22 crc kubenswrapper[4858]: I0320 09:33:22.355800 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8hv5z" podStartSLOduration=2.892405171 podStartE2EDuration="5.355777043s" podCreationTimestamp="2026-03-20 09:33:17 +0000 UTC" firstStartedPulling="2026-03-20 09:33:19.301512354 +0000 UTC m=+2180.621930551" lastFinishedPulling="2026-03-20 09:33:21.764884226 +0000 UTC m=+2183.085302423" observedRunningTime="2026-03-20 09:33:22.352572675 +0000 UTC m=+2183.672990872" watchObservedRunningTime="2026-03-20 09:33:22.355777043 +0000 UTC m=+2183.676195240" Mar 20 09:33:28 crc kubenswrapper[4858]: I0320 09:33:28.191550 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:28 crc kubenswrapper[4858]: I0320 09:33:28.192536 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:28 crc kubenswrapper[4858]: I0320 09:33:28.243731 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:28 crc kubenswrapper[4858]: I0320 09:33:28.497867 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:28 crc kubenswrapper[4858]: I0320 09:33:28.583697 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8hv5z"] Mar 20 09:33:30 crc kubenswrapper[4858]: I0320 09:33:30.406276 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8hv5z" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="registry-server" containerID="cri-o://77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1" gracePeriod=2 Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.292392 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.353597 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-catalog-content\") pod \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.416797 4858 generic.go:334] "Generic (PLEG): container finished" podID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerID="77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1" exitCode=0 Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.416856 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hv5z" event={"ID":"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f","Type":"ContainerDied","Data":"77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1"} Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.416892 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hv5z" event={"ID":"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f","Type":"ContainerDied","Data":"665ae1ffccb41796828550a6c79fff244d02c5573e2ce8648bcd4e06ab04f97e"} Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.416913 4858 scope.go:117] "RemoveContainer" containerID="77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.416915 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hv5z" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.441953 4858 scope.go:117] "RemoveContainer" containerID="78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.456187 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-utilities\") pod \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.456377 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfrzg\" (UniqueName: \"kubernetes.io/projected/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-kube-api-access-wfrzg\") pod \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\" (UID: \"0a8eb5d5-050e-42c0-8abe-3c4f5850b85f\") " Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.457304 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-utilities" (OuterVolumeSpecName: "utilities") pod "0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" (UID: "0a8eb5d5-050e-42c0-8abe-3c4f5850b85f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.464060 4858 scope.go:117] "RemoveContainer" containerID="ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.464196 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-kube-api-access-wfrzg" (OuterVolumeSpecName: "kube-api-access-wfrzg") pod "0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" (UID: "0a8eb5d5-050e-42c0-8abe-3c4f5850b85f"). InnerVolumeSpecName "kube-api-access-wfrzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.501749 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" (UID: "0a8eb5d5-050e-42c0-8abe-3c4f5850b85f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.519658 4858 scope.go:117] "RemoveContainer" containerID="77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1" Mar 20 09:33:31 crc kubenswrapper[4858]: E0320 09:33:31.520410 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1\": container with ID starting with 77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1 not found: ID does not exist" containerID="77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.520528 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1"} err="failed to get container status \"77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1\": rpc error: code = NotFound desc = could not find container \"77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1\": container with ID starting with 77a1e3a55a8f02878160aa962a9b3ae1a9cda6b080959ebbd51be84b8a7924b1 not found: ID does not exist" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.520609 4858 scope.go:117] "RemoveContainer" containerID="78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe" Mar 20 09:33:31 crc kubenswrapper[4858]: E0320 09:33:31.521428 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe\": container with ID starting with 78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe not found: ID does not exist" containerID="78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.521494 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe"} err="failed to get container status \"78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe\": rpc error: code = NotFound desc = could not find container \"78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe\": container with ID starting with 78eb6dd07f6d40e840ea09267b2e516bf60ec2c0841f75c34afb17228d7d51fe not found: ID does not exist" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.521544 4858 scope.go:117] "RemoveContainer" containerID="ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f" Mar 20 09:33:31 crc kubenswrapper[4858]: E0320 09:33:31.522153 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f\": container with ID starting with ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f not found: ID does not exist" containerID="ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.522285 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f"} err="failed to get container status \"ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f\": rpc error: code = NotFound desc = could not find container \"ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f\": container with ID starting with ede4392b01f331efd35622a300d118d6d2acc2d091e9d74573f2711d82aef38f not found: ID does not exist" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.558224 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.558644 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfrzg\" (UniqueName: \"kubernetes.io/projected/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-kube-api-access-wfrzg\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.558737 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.755791 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8hv5z"] Mar 20 09:33:31 crc kubenswrapper[4858]: I0320 09:33:31.763627 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8hv5z"] Mar 20 09:33:32 crc kubenswrapper[4858]: I0320 09:33:32.082553 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" path="/var/lib/kubelet/pods/0a8eb5d5-050e-42c0-8abe-3c4f5850b85f/volumes" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.145189 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566654-6ht6p"] Mar 20 09:34:00 crc kubenswrapper[4858]: E0320 09:34:00.152521 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="extract-content" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.152625 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="extract-content" Mar 20 09:34:00 crc kubenswrapper[4858]: E0320 09:34:00.152718 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.152778 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4858]: E0320 09:34:00.152853 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="extract-utilities" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.152916 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="extract-utilities" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.153154 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a8eb5d5-050e-42c0-8abe-3c4f5850b85f" containerName="registry-server" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.153891 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-6ht6p" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.156414 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.156726 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-6ht6p"] Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.160884 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.163848 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.249649 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92xnn\" (UniqueName: \"kubernetes.io/projected/d98bc030-f3d4-4753-b42b-f7674adc1816-kube-api-access-92xnn\") pod \"auto-csr-approver-29566654-6ht6p\" (UID: \"d98bc030-f3d4-4753-b42b-f7674adc1816\") " pod="openshift-infra/auto-csr-approver-29566654-6ht6p" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.351163 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92xnn\" (UniqueName: \"kubernetes.io/projected/d98bc030-f3d4-4753-b42b-f7674adc1816-kube-api-access-92xnn\") pod \"auto-csr-approver-29566654-6ht6p\" (UID: \"d98bc030-f3d4-4753-b42b-f7674adc1816\") " pod="openshift-infra/auto-csr-approver-29566654-6ht6p" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.376942 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92xnn\" (UniqueName: \"kubernetes.io/projected/d98bc030-f3d4-4753-b42b-f7674adc1816-kube-api-access-92xnn\") pod \"auto-csr-approver-29566654-6ht6p\" (UID: \"d98bc030-f3d4-4753-b42b-f7674adc1816\") " pod="openshift-infra/auto-csr-approver-29566654-6ht6p" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.475281 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-6ht6p" Mar 20 09:34:00 crc kubenswrapper[4858]: I0320 09:34:00.879422 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-6ht6p"] Mar 20 09:34:01 crc kubenswrapper[4858]: I0320 09:34:01.672732 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566654-6ht6p" event={"ID":"d98bc030-f3d4-4753-b42b-f7674adc1816","Type":"ContainerStarted","Data":"125edaacb1215b0ae7d632594a0122bc859b56e8286a1157fafbe27b030ebcd1"} Mar 20 09:34:03 crc kubenswrapper[4858]: I0320 09:34:03.692073 4858 generic.go:334] "Generic (PLEG): container finished" podID="d98bc030-f3d4-4753-b42b-f7674adc1816" containerID="c0003c9bd93e1176ada49158eb4c75ce389393064d2ed2e7bfa7638cd85f4fde" exitCode=0 Mar 20 09:34:03 crc kubenswrapper[4858]: I0320 09:34:03.692123 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566654-6ht6p" event={"ID":"d98bc030-f3d4-4753-b42b-f7674adc1816","Type":"ContainerDied","Data":"c0003c9bd93e1176ada49158eb4c75ce389393064d2ed2e7bfa7638cd85f4fde"} Mar 20 09:34:05 crc kubenswrapper[4858]: I0320 09:34:05.075403 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-6ht6p" Mar 20 09:34:05 crc kubenswrapper[4858]: I0320 09:34:05.128792 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92xnn\" (UniqueName: \"kubernetes.io/projected/d98bc030-f3d4-4753-b42b-f7674adc1816-kube-api-access-92xnn\") pod \"d98bc030-f3d4-4753-b42b-f7674adc1816\" (UID: \"d98bc030-f3d4-4753-b42b-f7674adc1816\") " Mar 20 09:34:05 crc kubenswrapper[4858]: I0320 09:34:05.143024 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d98bc030-f3d4-4753-b42b-f7674adc1816-kube-api-access-92xnn" (OuterVolumeSpecName: "kube-api-access-92xnn") pod "d98bc030-f3d4-4753-b42b-f7674adc1816" (UID: "d98bc030-f3d4-4753-b42b-f7674adc1816"). InnerVolumeSpecName "kube-api-access-92xnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:34:05 crc kubenswrapper[4858]: I0320 09:34:05.230953 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92xnn\" (UniqueName: \"kubernetes.io/projected/d98bc030-f3d4-4753-b42b-f7674adc1816-kube-api-access-92xnn\") on node \"crc\" DevicePath \"\"" Mar 20 09:34:05 crc kubenswrapper[4858]: I0320 09:34:05.708353 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566654-6ht6p" event={"ID":"d98bc030-f3d4-4753-b42b-f7674adc1816","Type":"ContainerDied","Data":"125edaacb1215b0ae7d632594a0122bc859b56e8286a1157fafbe27b030ebcd1"} Mar 20 09:34:05 crc kubenswrapper[4858]: I0320 09:34:05.708682 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="125edaacb1215b0ae7d632594a0122bc859b56e8286a1157fafbe27b030ebcd1" Mar 20 09:34:05 crc kubenswrapper[4858]: I0320 09:34:05.708415 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566654-6ht6p" Mar 20 09:34:06 crc kubenswrapper[4858]: I0320 09:34:06.156758 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-dc294"] Mar 20 09:34:06 crc kubenswrapper[4858]: I0320 09:34:06.164420 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566648-dc294"] Mar 20 09:34:07 crc kubenswrapper[4858]: I0320 09:34:07.890361 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:34:07 crc kubenswrapper[4858]: I0320 09:34:07.890973 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:34:08 crc kubenswrapper[4858]: I0320 09:34:08.080650 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f063b8c5-82cd-4d57-96f5-b91adc5717b2" path="/var/lib/kubelet/pods/f063b8c5-82cd-4d57-96f5-b91adc5717b2/volumes" Mar 20 09:34:30 crc kubenswrapper[4858]: I0320 09:34:30.914139 4858 scope.go:117] "RemoveContainer" containerID="686926f47c91f575054458d02df1066f27f634331ffea7a6d74facb2a5c63e46" Mar 20 09:34:37 crc kubenswrapper[4858]: I0320 09:34:37.890588 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:34:37 crc kubenswrapper[4858]: I0320 09:34:37.891270 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.462617 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tzhg7"] Mar 20 09:34:53 crc kubenswrapper[4858]: E0320 09:34:53.464203 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d98bc030-f3d4-4753-b42b-f7674adc1816" containerName="oc" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.464224 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="d98bc030-f3d4-4753-b42b-f7674adc1816" containerName="oc" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.464406 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="d98bc030-f3d4-4753-b42b-f7674adc1816" containerName="oc" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.465872 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.479046 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tzhg7"] Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.647814 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jt8nr\" (UniqueName: \"kubernetes.io/projected/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-kube-api-access-jt8nr\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.647868 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-utilities\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.647896 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-catalog-content\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.748934 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jt8nr\" (UniqueName: \"kubernetes.io/projected/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-kube-api-access-jt8nr\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.748987 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-utilities\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.749010 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-catalog-content\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.749497 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-catalog-content\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.749581 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-utilities\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.769261 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jt8nr\" (UniqueName: \"kubernetes.io/projected/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-kube-api-access-jt8nr\") pod \"community-operators-tzhg7\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:53 crc kubenswrapper[4858]: I0320 09:34:53.788919 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:34:54 crc kubenswrapper[4858]: I0320 09:34:54.310611 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tzhg7"] Mar 20 09:34:55 crc kubenswrapper[4858]: I0320 09:34:55.109866 4858 generic.go:334] "Generic (PLEG): container finished" podID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerID="5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34" exitCode=0 Mar 20 09:34:55 crc kubenswrapper[4858]: I0320 09:34:55.109951 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tzhg7" event={"ID":"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21","Type":"ContainerDied","Data":"5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34"} Mar 20 09:34:55 crc kubenswrapper[4858]: I0320 09:34:55.110177 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tzhg7" event={"ID":"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21","Type":"ContainerStarted","Data":"6fc7ce623c3fa6de5bf295c971565005ced408db61551ce51c12439b86c4bd13"} Mar 20 09:34:56 crc kubenswrapper[4858]: I0320 09:34:56.119028 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tzhg7" event={"ID":"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21","Type":"ContainerStarted","Data":"5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f"} Mar 20 09:34:56 crc kubenswrapper[4858]: E0320 09:34:56.161661 4858 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95db68a3_72b5_46e0_a6fa_2a32cd6b6a21.slice/crio-5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f.scope\": RecentStats: unable to find data in memory cache]" Mar 20 09:34:57 crc kubenswrapper[4858]: I0320 09:34:57.129344 4858 generic.go:334] "Generic (PLEG): container finished" podID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerID="5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f" exitCode=0 Mar 20 09:34:57 crc kubenswrapper[4858]: I0320 09:34:57.129454 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tzhg7" event={"ID":"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21","Type":"ContainerDied","Data":"5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f"} Mar 20 09:34:59 crc kubenswrapper[4858]: I0320 09:34:59.148075 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tzhg7" event={"ID":"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21","Type":"ContainerStarted","Data":"13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b"} Mar 20 09:34:59 crc kubenswrapper[4858]: I0320 09:34:59.172598 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tzhg7" podStartSLOduration=2.627144203 podStartE2EDuration="6.17257539s" podCreationTimestamp="2026-03-20 09:34:53 +0000 UTC" firstStartedPulling="2026-03-20 09:34:55.112347579 +0000 UTC m=+2276.432765776" lastFinishedPulling="2026-03-20 09:34:58.657778766 +0000 UTC m=+2279.978196963" observedRunningTime="2026-03-20 09:34:59.167174993 +0000 UTC m=+2280.487593190" watchObservedRunningTime="2026-03-20 09:34:59.17257539 +0000 UTC m=+2280.492993587" Mar 20 09:35:03 crc kubenswrapper[4858]: I0320 09:35:03.789799 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:35:03 crc kubenswrapper[4858]: I0320 09:35:03.790199 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:35:03 crc kubenswrapper[4858]: I0320 09:35:03.837966 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:35:04 crc kubenswrapper[4858]: I0320 09:35:04.242918 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:35:04 crc kubenswrapper[4858]: I0320 09:35:04.339403 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tzhg7"] Mar 20 09:35:06 crc kubenswrapper[4858]: I0320 09:35:06.191861 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tzhg7" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="registry-server" containerID="cri-o://13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b" gracePeriod=2 Mar 20 09:35:06 crc kubenswrapper[4858]: E0320 09:35:06.365464 4858 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95db68a3_72b5_46e0_a6fa_2a32cd6b6a21.slice/crio-13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b.scope\": RecentStats: unable to find data in memory cache]" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.481348 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pn8c8"] Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.483285 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.494811 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pn8c8"] Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.554444 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-utilities\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.554517 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-catalog-content\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.554655 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p4fl\" (UniqueName: \"kubernetes.io/projected/e604f326-5eb0-480a-936a-cdc55e47f1f2-kube-api-access-9p4fl\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.657659 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p4fl\" (UniqueName: \"kubernetes.io/projected/e604f326-5eb0-480a-936a-cdc55e47f1f2-kube-api-access-9p4fl\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.658511 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-utilities\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.659122 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-utilities\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.659640 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-catalog-content\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.659985 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-catalog-content\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.679757 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p4fl\" (UniqueName: \"kubernetes.io/projected/e604f326-5eb0-480a-936a-cdc55e47f1f2-kube-api-access-9p4fl\") pod \"certified-operators-pn8c8\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.808665 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.855255 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.889984 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.890118 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.890201 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.890834 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5817ed13dcb1fb797014d77cda07222984a083879566a3f0d5db8e9e973bd960"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.890894 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://5817ed13dcb1fb797014d77cda07222984a083879566a3f0d5db8e9e973bd960" gracePeriod=600 Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.965195 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-utilities\") pod \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.965340 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-catalog-content\") pod \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.965465 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jt8nr\" (UniqueName: \"kubernetes.io/projected/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-kube-api-access-jt8nr\") pod \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\" (UID: \"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21\") " Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.966176 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-utilities" (OuterVolumeSpecName: "utilities") pod "95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" (UID: "95db68a3-72b5-46e0-a6fa-2a32cd6b6a21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:35:07 crc kubenswrapper[4858]: I0320 09:35:07.977865 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-kube-api-access-jt8nr" (OuterVolumeSpecName: "kube-api-access-jt8nr") pod "95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" (UID: "95db68a3-72b5-46e0-a6fa-2a32cd6b6a21"). InnerVolumeSpecName "kube-api-access-jt8nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.028179 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" (UID: "95db68a3-72b5-46e0-a6fa-2a32cd6b6a21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.069297 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.069380 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jt8nr\" (UniqueName: \"kubernetes.io/projected/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-kube-api-access-jt8nr\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.069395 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.144302 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pn8c8"] Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.223437 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="5817ed13dcb1fb797014d77cda07222984a083879566a3f0d5db8e9e973bd960" exitCode=0 Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.223641 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"5817ed13dcb1fb797014d77cda07222984a083879566a3f0d5db8e9e973bd960"} Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.223933 4858 scope.go:117] "RemoveContainer" containerID="f458c5b3b527e2cc1a5fe69c7543069ef362d12c40106fb3284464ab05451cea" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.233460 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn8c8" event={"ID":"e604f326-5eb0-480a-936a-cdc55e47f1f2","Type":"ContainerStarted","Data":"271b3c9729608f0985f805649b9cd638edf17acaaf8b76b5592f8f593bf7c7cc"} Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.252521 4858 generic.go:334] "Generic (PLEG): container finished" podID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerID="13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b" exitCode=0 Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.252603 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tzhg7" event={"ID":"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21","Type":"ContainerDied","Data":"13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b"} Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.252633 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tzhg7" event={"ID":"95db68a3-72b5-46e0-a6fa-2a32cd6b6a21","Type":"ContainerDied","Data":"6fc7ce623c3fa6de5bf295c971565005ced408db61551ce51c12439b86c4bd13"} Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.252696 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tzhg7" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.263086 4858 scope.go:117] "RemoveContainer" containerID="13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.287392 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tzhg7"] Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.293439 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tzhg7"] Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.307902 4858 scope.go:117] "RemoveContainer" containerID="5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.329756 4858 scope.go:117] "RemoveContainer" containerID="5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.397493 4858 scope.go:117] "RemoveContainer" containerID="13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b" Mar 20 09:35:08 crc kubenswrapper[4858]: E0320 09:35:08.398004 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b\": container with ID starting with 13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b not found: ID does not exist" containerID="13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.398045 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b"} err="failed to get container status \"13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b\": rpc error: code = NotFound desc = could not find container \"13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b\": container with ID starting with 13557fda7073506a4978a22aba64db3d3c9bc723026764ef9c5ad2ffb1ab323b not found: ID does not exist" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.398254 4858 scope.go:117] "RemoveContainer" containerID="5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f" Mar 20 09:35:08 crc kubenswrapper[4858]: E0320 09:35:08.398820 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f\": container with ID starting with 5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f not found: ID does not exist" containerID="5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.398862 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f"} err="failed to get container status \"5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f\": rpc error: code = NotFound desc = could not find container \"5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f\": container with ID starting with 5417d7101df766fdaef164a44994a2ad270ae602fca5bf4d526efeca67afcc6f not found: ID does not exist" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.398889 4858 scope.go:117] "RemoveContainer" containerID="5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34" Mar 20 09:35:08 crc kubenswrapper[4858]: E0320 09:35:08.399173 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34\": container with ID starting with 5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34 not found: ID does not exist" containerID="5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34" Mar 20 09:35:08 crc kubenswrapper[4858]: I0320 09:35:08.399253 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34"} err="failed to get container status \"5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34\": rpc error: code = NotFound desc = could not find container \"5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34\": container with ID starting with 5393f0ace69b2385deb5d0d9ce36f92034edc1ac1057353ebbdeb246560ecb34 not found: ID does not exist" Mar 20 09:35:09 crc kubenswrapper[4858]: I0320 09:35:09.266777 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6"} Mar 20 09:35:09 crc kubenswrapper[4858]: I0320 09:35:09.268414 4858 generic.go:334] "Generic (PLEG): container finished" podID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerID="0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818" exitCode=0 Mar 20 09:35:09 crc kubenswrapper[4858]: I0320 09:35:09.268504 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn8c8" event={"ID":"e604f326-5eb0-480a-936a-cdc55e47f1f2","Type":"ContainerDied","Data":"0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818"} Mar 20 09:35:09 crc kubenswrapper[4858]: I0320 09:35:09.271232 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:35:10 crc kubenswrapper[4858]: I0320 09:35:10.099298 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" path="/var/lib/kubelet/pods/95db68a3-72b5-46e0-a6fa-2a32cd6b6a21/volumes" Mar 20 09:35:10 crc kubenswrapper[4858]: I0320 09:35:10.278946 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn8c8" event={"ID":"e604f326-5eb0-480a-936a-cdc55e47f1f2","Type":"ContainerStarted","Data":"34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7"} Mar 20 09:35:11 crc kubenswrapper[4858]: I0320 09:35:11.291364 4858 generic.go:334] "Generic (PLEG): container finished" podID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerID="34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7" exitCode=0 Mar 20 09:35:11 crc kubenswrapper[4858]: I0320 09:35:11.291445 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn8c8" event={"ID":"e604f326-5eb0-480a-936a-cdc55e47f1f2","Type":"ContainerDied","Data":"34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7"} Mar 20 09:35:12 crc kubenswrapper[4858]: I0320 09:35:12.304516 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn8c8" event={"ID":"e604f326-5eb0-480a-936a-cdc55e47f1f2","Type":"ContainerStarted","Data":"e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7"} Mar 20 09:35:12 crc kubenswrapper[4858]: I0320 09:35:12.325057 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pn8c8" podStartSLOduration=2.736482121 podStartE2EDuration="5.325039817s" podCreationTimestamp="2026-03-20 09:35:07 +0000 UTC" firstStartedPulling="2026-03-20 09:35:09.27095393 +0000 UTC m=+2290.591372127" lastFinishedPulling="2026-03-20 09:35:11.859511626 +0000 UTC m=+2293.179929823" observedRunningTime="2026-03-20 09:35:12.324916233 +0000 UTC m=+2293.645334440" watchObservedRunningTime="2026-03-20 09:35:12.325039817 +0000 UTC m=+2293.645458014" Mar 20 09:35:17 crc kubenswrapper[4858]: I0320 09:35:17.856508 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:17 crc kubenswrapper[4858]: I0320 09:35:17.859857 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:17 crc kubenswrapper[4858]: I0320 09:35:17.911783 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:18 crc kubenswrapper[4858]: I0320 09:35:18.404813 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:18 crc kubenswrapper[4858]: I0320 09:35:18.456306 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pn8c8"] Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.372877 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pn8c8" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="registry-server" containerID="cri-o://e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7" gracePeriod=2 Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.772293 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.870365 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-utilities\") pod \"e604f326-5eb0-480a-936a-cdc55e47f1f2\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.870524 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p4fl\" (UniqueName: \"kubernetes.io/projected/e604f326-5eb0-480a-936a-cdc55e47f1f2-kube-api-access-9p4fl\") pod \"e604f326-5eb0-480a-936a-cdc55e47f1f2\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.870842 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-catalog-content\") pod \"e604f326-5eb0-480a-936a-cdc55e47f1f2\" (UID: \"e604f326-5eb0-480a-936a-cdc55e47f1f2\") " Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.871638 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-utilities" (OuterVolumeSpecName: "utilities") pod "e604f326-5eb0-480a-936a-cdc55e47f1f2" (UID: "e604f326-5eb0-480a-936a-cdc55e47f1f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.878115 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e604f326-5eb0-480a-936a-cdc55e47f1f2-kube-api-access-9p4fl" (OuterVolumeSpecName: "kube-api-access-9p4fl") pod "e604f326-5eb0-480a-936a-cdc55e47f1f2" (UID: "e604f326-5eb0-480a-936a-cdc55e47f1f2"). InnerVolumeSpecName "kube-api-access-9p4fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.972861 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:20 crc kubenswrapper[4858]: I0320 09:35:20.972898 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p4fl\" (UniqueName: \"kubernetes.io/projected/e604f326-5eb0-480a-936a-cdc55e47f1f2-kube-api-access-9p4fl\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.217481 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e604f326-5eb0-480a-936a-cdc55e47f1f2" (UID: "e604f326-5eb0-480a-936a-cdc55e47f1f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.276913 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e604f326-5eb0-480a-936a-cdc55e47f1f2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.384972 4858 generic.go:334] "Generic (PLEG): container finished" podID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerID="e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7" exitCode=0 Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.385095 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn8c8" event={"ID":"e604f326-5eb0-480a-936a-cdc55e47f1f2","Type":"ContainerDied","Data":"e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7"} Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.385232 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pn8c8" event={"ID":"e604f326-5eb0-480a-936a-cdc55e47f1f2","Type":"ContainerDied","Data":"271b3c9729608f0985f805649b9cd638edf17acaaf8b76b5592f8f593bf7c7cc"} Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.385347 4858 scope.go:117] "RemoveContainer" containerID="e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.385501 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pn8c8" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.404416 4858 scope.go:117] "RemoveContainer" containerID="34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.430150 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pn8c8"] Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.435571 4858 scope.go:117] "RemoveContainer" containerID="0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.438429 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pn8c8"] Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.469646 4858 scope.go:117] "RemoveContainer" containerID="e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7" Mar 20 09:35:21 crc kubenswrapper[4858]: E0320 09:35:21.471647 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7\": container with ID starting with e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7 not found: ID does not exist" containerID="e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.471707 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7"} err="failed to get container status \"e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7\": rpc error: code = NotFound desc = could not find container \"e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7\": container with ID starting with e7c73a167222fba9eedb66248ace8d2df458fcb78cfc28cdeea4d0a25511b5d7 not found: ID does not exist" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.471740 4858 scope.go:117] "RemoveContainer" containerID="34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7" Mar 20 09:35:21 crc kubenswrapper[4858]: E0320 09:35:21.472063 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7\": container with ID starting with 34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7 not found: ID does not exist" containerID="34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.472110 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7"} err="failed to get container status \"34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7\": rpc error: code = NotFound desc = could not find container \"34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7\": container with ID starting with 34651562bb364acdf5b40888acc8807874e622ea5c803e559c71a995631457e7 not found: ID does not exist" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.472145 4858 scope.go:117] "RemoveContainer" containerID="0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818" Mar 20 09:35:21 crc kubenswrapper[4858]: E0320 09:35:21.472489 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818\": container with ID starting with 0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818 not found: ID does not exist" containerID="0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818" Mar 20 09:35:21 crc kubenswrapper[4858]: I0320 09:35:21.472541 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818"} err="failed to get container status \"0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818\": rpc error: code = NotFound desc = could not find container \"0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818\": container with ID starting with 0fd52a705be87afd6b3b31b09ca2107755fe6017020ba18acf613921d513c818 not found: ID does not exist" Mar 20 09:35:22 crc kubenswrapper[4858]: I0320 09:35:22.081438 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" path="/var/lib/kubelet/pods/e604f326-5eb0-480a-936a-cdc55e47f1f2/volumes" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.161811 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566656-ntc4k"] Mar 20 09:36:00 crc kubenswrapper[4858]: E0320 09:36:00.162810 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.162826 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4858]: E0320 09:36:00.162846 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.162855 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4858]: E0320 09:36:00.162863 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="extract-content" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.162872 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="extract-content" Mar 20 09:36:00 crc kubenswrapper[4858]: E0320 09:36:00.162893 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="extract-content" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.162901 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="extract-content" Mar 20 09:36:00 crc kubenswrapper[4858]: E0320 09:36:00.162916 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="extract-utilities" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.162927 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="extract-utilities" Mar 20 09:36:00 crc kubenswrapper[4858]: E0320 09:36:00.162942 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="extract-utilities" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.162950 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="extract-utilities" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.163118 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="e604f326-5eb0-480a-936a-cdc55e47f1f2" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.163142 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="95db68a3-72b5-46e0-a6fa-2a32cd6b6a21" containerName="registry-server" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.163729 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-ntc4k" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.167298 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.167724 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.167908 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.178045 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-ntc4k"] Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.308180 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9585\" (UniqueName: \"kubernetes.io/projected/08ff6909-89ba-4b73-9940-eb06289f2a30-kube-api-access-p9585\") pod \"auto-csr-approver-29566656-ntc4k\" (UID: \"08ff6909-89ba-4b73-9940-eb06289f2a30\") " pod="openshift-infra/auto-csr-approver-29566656-ntc4k" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.409881 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9585\" (UniqueName: \"kubernetes.io/projected/08ff6909-89ba-4b73-9940-eb06289f2a30-kube-api-access-p9585\") pod \"auto-csr-approver-29566656-ntc4k\" (UID: \"08ff6909-89ba-4b73-9940-eb06289f2a30\") " pod="openshift-infra/auto-csr-approver-29566656-ntc4k" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.434032 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9585\" (UniqueName: \"kubernetes.io/projected/08ff6909-89ba-4b73-9940-eb06289f2a30-kube-api-access-p9585\") pod \"auto-csr-approver-29566656-ntc4k\" (UID: \"08ff6909-89ba-4b73-9940-eb06289f2a30\") " pod="openshift-infra/auto-csr-approver-29566656-ntc4k" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.488165 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-ntc4k" Mar 20 09:36:00 crc kubenswrapper[4858]: I0320 09:36:00.935062 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-ntc4k"] Mar 20 09:36:01 crc kubenswrapper[4858]: I0320 09:36:01.715398 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566656-ntc4k" event={"ID":"08ff6909-89ba-4b73-9940-eb06289f2a30","Type":"ContainerStarted","Data":"51b359ca05c2ca89b56886ed5b42311f7c423eb84b25b4841d534ab684105c55"} Mar 20 09:36:03 crc kubenswrapper[4858]: I0320 09:36:03.729921 4858 generic.go:334] "Generic (PLEG): container finished" podID="08ff6909-89ba-4b73-9940-eb06289f2a30" containerID="101c55c930a879083435516e891376919f72cbb9e76b90005e15b16685d68814" exitCode=0 Mar 20 09:36:03 crc kubenswrapper[4858]: I0320 09:36:03.729972 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566656-ntc4k" event={"ID":"08ff6909-89ba-4b73-9940-eb06289f2a30","Type":"ContainerDied","Data":"101c55c930a879083435516e891376919f72cbb9e76b90005e15b16685d68814"} Mar 20 09:36:05 crc kubenswrapper[4858]: I0320 09:36:05.048792 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-ntc4k" Mar 20 09:36:05 crc kubenswrapper[4858]: I0320 09:36:05.189135 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9585\" (UniqueName: \"kubernetes.io/projected/08ff6909-89ba-4b73-9940-eb06289f2a30-kube-api-access-p9585\") pod \"08ff6909-89ba-4b73-9940-eb06289f2a30\" (UID: \"08ff6909-89ba-4b73-9940-eb06289f2a30\") " Mar 20 09:36:05 crc kubenswrapper[4858]: I0320 09:36:05.196475 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ff6909-89ba-4b73-9940-eb06289f2a30-kube-api-access-p9585" (OuterVolumeSpecName: "kube-api-access-p9585") pod "08ff6909-89ba-4b73-9940-eb06289f2a30" (UID: "08ff6909-89ba-4b73-9940-eb06289f2a30"). InnerVolumeSpecName "kube-api-access-p9585". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:36:05 crc kubenswrapper[4858]: I0320 09:36:05.291168 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9585\" (UniqueName: \"kubernetes.io/projected/08ff6909-89ba-4b73-9940-eb06289f2a30-kube-api-access-p9585\") on node \"crc\" DevicePath \"\"" Mar 20 09:36:05 crc kubenswrapper[4858]: I0320 09:36:05.747799 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566656-ntc4k" event={"ID":"08ff6909-89ba-4b73-9940-eb06289f2a30","Type":"ContainerDied","Data":"51b359ca05c2ca89b56886ed5b42311f7c423eb84b25b4841d534ab684105c55"} Mar 20 09:36:05 crc kubenswrapper[4858]: I0320 09:36:05.747855 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51b359ca05c2ca89b56886ed5b42311f7c423eb84b25b4841d534ab684105c55" Mar 20 09:36:05 crc kubenswrapper[4858]: I0320 09:36:05.747874 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566656-ntc4k" Mar 20 09:36:06 crc kubenswrapper[4858]: I0320 09:36:06.135803 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-prn7d"] Mar 20 09:36:06 crc kubenswrapper[4858]: I0320 09:36:06.142125 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566650-prn7d"] Mar 20 09:36:08 crc kubenswrapper[4858]: I0320 09:36:08.079953 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61c56a2-c37b-4b49-b970-0da0a98defcc" path="/var/lib/kubelet/pods/c61c56a2-c37b-4b49-b970-0da0a98defcc/volumes" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.135908 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pmkq7"] Mar 20 09:36:21 crc kubenswrapper[4858]: E0320 09:36:21.141586 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ff6909-89ba-4b73-9940-eb06289f2a30" containerName="oc" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.141733 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ff6909-89ba-4b73-9940-eb06289f2a30" containerName="oc" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.142015 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ff6909-89ba-4b73-9940-eb06289f2a30" containerName="oc" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.143674 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.170459 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pmkq7"] Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.247185 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-utilities\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.247345 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-catalog-content\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.247424 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfrbx\" (UniqueName: \"kubernetes.io/projected/9f37301c-6b62-4467-b85a-40560ee0071d-kube-api-access-kfrbx\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.349306 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfrbx\" (UniqueName: \"kubernetes.io/projected/9f37301c-6b62-4467-b85a-40560ee0071d-kube-api-access-kfrbx\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.349429 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-utilities\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.349541 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-catalog-content\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.350384 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-catalog-content\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.350428 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-utilities\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.374861 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfrbx\" (UniqueName: \"kubernetes.io/projected/9f37301c-6b62-4467-b85a-40560ee0071d-kube-api-access-kfrbx\") pod \"redhat-marketplace-pmkq7\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.476501 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:21 crc kubenswrapper[4858]: I0320 09:36:21.936024 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pmkq7"] Mar 20 09:36:22 crc kubenswrapper[4858]: I0320 09:36:22.890932 4858 generic.go:334] "Generic (PLEG): container finished" podID="9f37301c-6b62-4467-b85a-40560ee0071d" containerID="0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7" exitCode=0 Mar 20 09:36:22 crc kubenswrapper[4858]: I0320 09:36:22.891010 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pmkq7" event={"ID":"9f37301c-6b62-4467-b85a-40560ee0071d","Type":"ContainerDied","Data":"0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7"} Mar 20 09:36:22 crc kubenswrapper[4858]: I0320 09:36:22.891371 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pmkq7" event={"ID":"9f37301c-6b62-4467-b85a-40560ee0071d","Type":"ContainerStarted","Data":"e1d3bdea9724b7cfbdb6e8d506aaa9fb679d24db3302a634d210db1bb2130a9b"} Mar 20 09:36:24 crc kubenswrapper[4858]: I0320 09:36:24.913757 4858 generic.go:334] "Generic (PLEG): container finished" podID="9f37301c-6b62-4467-b85a-40560ee0071d" containerID="d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b" exitCode=0 Mar 20 09:36:24 crc kubenswrapper[4858]: I0320 09:36:24.913923 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pmkq7" event={"ID":"9f37301c-6b62-4467-b85a-40560ee0071d","Type":"ContainerDied","Data":"d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b"} Mar 20 09:36:25 crc kubenswrapper[4858]: I0320 09:36:25.925137 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pmkq7" event={"ID":"9f37301c-6b62-4467-b85a-40560ee0071d","Type":"ContainerStarted","Data":"11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae"} Mar 20 09:36:25 crc kubenswrapper[4858]: I0320 09:36:25.946697 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pmkq7" podStartSLOduration=2.481488972 podStartE2EDuration="4.946670339s" podCreationTimestamp="2026-03-20 09:36:21 +0000 UTC" firstStartedPulling="2026-03-20 09:36:22.896544026 +0000 UTC m=+2364.216962223" lastFinishedPulling="2026-03-20 09:36:25.361725393 +0000 UTC m=+2366.682143590" observedRunningTime="2026-03-20 09:36:25.943555238 +0000 UTC m=+2367.263973455" watchObservedRunningTime="2026-03-20 09:36:25.946670339 +0000 UTC m=+2367.267088536" Mar 20 09:36:31 crc kubenswrapper[4858]: I0320 09:36:31.094415 4858 scope.go:117] "RemoveContainer" containerID="4f40169bac1ee45370466ab61edba74c94ad2f30f2ed5563f57e2a6a2fa2c48b" Mar 20 09:36:31 crc kubenswrapper[4858]: I0320 09:36:31.477466 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:31 crc kubenswrapper[4858]: I0320 09:36:31.478070 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:31 crc kubenswrapper[4858]: I0320 09:36:31.521048 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:32 crc kubenswrapper[4858]: I0320 09:36:32.041737 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:32 crc kubenswrapper[4858]: I0320 09:36:32.113810 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pmkq7"] Mar 20 09:36:33 crc kubenswrapper[4858]: I0320 09:36:33.987808 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pmkq7" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="registry-server" containerID="cri-o://11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae" gracePeriod=2 Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.438626 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.569174 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-catalog-content\") pod \"9f37301c-6b62-4467-b85a-40560ee0071d\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.569298 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfrbx\" (UniqueName: \"kubernetes.io/projected/9f37301c-6b62-4467-b85a-40560ee0071d-kube-api-access-kfrbx\") pod \"9f37301c-6b62-4467-b85a-40560ee0071d\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.569412 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-utilities\") pod \"9f37301c-6b62-4467-b85a-40560ee0071d\" (UID: \"9f37301c-6b62-4467-b85a-40560ee0071d\") " Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.570509 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-utilities" (OuterVolumeSpecName: "utilities") pod "9f37301c-6b62-4467-b85a-40560ee0071d" (UID: "9f37301c-6b62-4467-b85a-40560ee0071d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.574386 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f37301c-6b62-4467-b85a-40560ee0071d-kube-api-access-kfrbx" (OuterVolumeSpecName: "kube-api-access-kfrbx") pod "9f37301c-6b62-4467-b85a-40560ee0071d" (UID: "9f37301c-6b62-4467-b85a-40560ee0071d"). InnerVolumeSpecName "kube-api-access-kfrbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.594384 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f37301c-6b62-4467-b85a-40560ee0071d" (UID: "9f37301c-6b62-4467-b85a-40560ee0071d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.671741 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.671809 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f37301c-6b62-4467-b85a-40560ee0071d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.671827 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfrbx\" (UniqueName: \"kubernetes.io/projected/9f37301c-6b62-4467-b85a-40560ee0071d-kube-api-access-kfrbx\") on node \"crc\" DevicePath \"\"" Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.999769 4858 generic.go:334] "Generic (PLEG): container finished" podID="9f37301c-6b62-4467-b85a-40560ee0071d" containerID="11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae" exitCode=0 Mar 20 09:36:34 crc kubenswrapper[4858]: I0320 09:36:34.999825 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pmkq7" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:34.999817 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pmkq7" event={"ID":"9f37301c-6b62-4467-b85a-40560ee0071d","Type":"ContainerDied","Data":"11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae"} Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:34.999976 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pmkq7" event={"ID":"9f37301c-6b62-4467-b85a-40560ee0071d","Type":"ContainerDied","Data":"e1d3bdea9724b7cfbdb6e8d506aaa9fb679d24db3302a634d210db1bb2130a9b"} Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.000000 4858 scope.go:117] "RemoveContainer" containerID="11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.031229 4858 scope.go:117] "RemoveContainer" containerID="d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.031391 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pmkq7"] Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.040206 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pmkq7"] Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.058503 4858 scope.go:117] "RemoveContainer" containerID="0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.075653 4858 scope.go:117] "RemoveContainer" containerID="11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae" Mar 20 09:36:35 crc kubenswrapper[4858]: E0320 09:36:35.076126 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae\": container with ID starting with 11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae not found: ID does not exist" containerID="11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.076193 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae"} err="failed to get container status \"11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae\": rpc error: code = NotFound desc = could not find container \"11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae\": container with ID starting with 11a20daa517fdf980098e63c5c40277cc0237707a17f81d132dbe477f549b7ae not found: ID does not exist" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.076236 4858 scope.go:117] "RemoveContainer" containerID="d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b" Mar 20 09:36:35 crc kubenswrapper[4858]: E0320 09:36:35.076633 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b\": container with ID starting with d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b not found: ID does not exist" containerID="d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.076687 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b"} err="failed to get container status \"d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b\": rpc error: code = NotFound desc = could not find container \"d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b\": container with ID starting with d7f76f53c5f0d643fc5d8617265ed059cd558ebe0526ab7247ad9030a3dede9b not found: ID does not exist" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.076723 4858 scope.go:117] "RemoveContainer" containerID="0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7" Mar 20 09:36:35 crc kubenswrapper[4858]: E0320 09:36:35.076996 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7\": container with ID starting with 0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7 not found: ID does not exist" containerID="0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7" Mar 20 09:36:35 crc kubenswrapper[4858]: I0320 09:36:35.077047 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7"} err="failed to get container status \"0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7\": rpc error: code = NotFound desc = could not find container \"0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7\": container with ID starting with 0feb300d88536045469547fab1d2bc2054c49c36e5cf9f57437393320b8bfcc7 not found: ID does not exist" Mar 20 09:36:36 crc kubenswrapper[4858]: I0320 09:36:36.078011 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" path="/var/lib/kubelet/pods/9f37301c-6b62-4467-b85a-40560ee0071d/volumes" Mar 20 09:37:37 crc kubenswrapper[4858]: I0320 09:37:37.890665 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:37:37 crc kubenswrapper[4858]: I0320 09:37:37.891243 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.148585 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566658-tht4s"] Mar 20 09:38:00 crc kubenswrapper[4858]: E0320 09:38:00.149469 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="extract-content" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.149482 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="extract-content" Mar 20 09:38:00 crc kubenswrapper[4858]: E0320 09:38:00.149503 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="extract-utilities" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.149511 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="extract-utilities" Mar 20 09:38:00 crc kubenswrapper[4858]: E0320 09:38:00.149523 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="registry-server" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.149528 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="registry-server" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.149658 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f37301c-6b62-4467-b85a-40560ee0071d" containerName="registry-server" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.150162 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-tht4s" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.152160 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.153059 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.153179 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.165139 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-tht4s"] Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.272719 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkxg7\" (UniqueName: \"kubernetes.io/projected/1daa8fc3-43fa-4b60-a94a-4594306737d5-kube-api-access-bkxg7\") pod \"auto-csr-approver-29566658-tht4s\" (UID: \"1daa8fc3-43fa-4b60-a94a-4594306737d5\") " pod="openshift-infra/auto-csr-approver-29566658-tht4s" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.374733 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkxg7\" (UniqueName: \"kubernetes.io/projected/1daa8fc3-43fa-4b60-a94a-4594306737d5-kube-api-access-bkxg7\") pod \"auto-csr-approver-29566658-tht4s\" (UID: \"1daa8fc3-43fa-4b60-a94a-4594306737d5\") " pod="openshift-infra/auto-csr-approver-29566658-tht4s" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.395174 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkxg7\" (UniqueName: \"kubernetes.io/projected/1daa8fc3-43fa-4b60-a94a-4594306737d5-kube-api-access-bkxg7\") pod \"auto-csr-approver-29566658-tht4s\" (UID: \"1daa8fc3-43fa-4b60-a94a-4594306737d5\") " pod="openshift-infra/auto-csr-approver-29566658-tht4s" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.530461 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-tht4s" Mar 20 09:38:00 crc kubenswrapper[4858]: I0320 09:38:00.984733 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-tht4s"] Mar 20 09:38:01 crc kubenswrapper[4858]: I0320 09:38:01.694972 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566658-tht4s" event={"ID":"1daa8fc3-43fa-4b60-a94a-4594306737d5","Type":"ContainerStarted","Data":"afc14f30d7a66a863c473afd0f6615629602bb7edbb3434dd8a526a12fd24177"} Mar 20 09:38:02 crc kubenswrapper[4858]: I0320 09:38:02.702843 4858 generic.go:334] "Generic (PLEG): container finished" podID="1daa8fc3-43fa-4b60-a94a-4594306737d5" containerID="62fa5939ee5b6f50beb2399d7ce0432f46efdffdda67784869b7fb84502a48fd" exitCode=0 Mar 20 09:38:02 crc kubenswrapper[4858]: I0320 09:38:02.702934 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566658-tht4s" event={"ID":"1daa8fc3-43fa-4b60-a94a-4594306737d5","Type":"ContainerDied","Data":"62fa5939ee5b6f50beb2399d7ce0432f46efdffdda67784869b7fb84502a48fd"} Mar 20 09:38:04 crc kubenswrapper[4858]: I0320 09:38:04.044900 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-tht4s" Mar 20 09:38:04 crc kubenswrapper[4858]: I0320 09:38:04.140301 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkxg7\" (UniqueName: \"kubernetes.io/projected/1daa8fc3-43fa-4b60-a94a-4594306737d5-kube-api-access-bkxg7\") pod \"1daa8fc3-43fa-4b60-a94a-4594306737d5\" (UID: \"1daa8fc3-43fa-4b60-a94a-4594306737d5\") " Mar 20 09:38:04 crc kubenswrapper[4858]: I0320 09:38:04.146897 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1daa8fc3-43fa-4b60-a94a-4594306737d5-kube-api-access-bkxg7" (OuterVolumeSpecName: "kube-api-access-bkxg7") pod "1daa8fc3-43fa-4b60-a94a-4594306737d5" (UID: "1daa8fc3-43fa-4b60-a94a-4594306737d5"). InnerVolumeSpecName "kube-api-access-bkxg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:38:04 crc kubenswrapper[4858]: I0320 09:38:04.242671 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkxg7\" (UniqueName: \"kubernetes.io/projected/1daa8fc3-43fa-4b60-a94a-4594306737d5-kube-api-access-bkxg7\") on node \"crc\" DevicePath \"\"" Mar 20 09:38:04 crc kubenswrapper[4858]: I0320 09:38:04.720968 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566658-tht4s" event={"ID":"1daa8fc3-43fa-4b60-a94a-4594306737d5","Type":"ContainerDied","Data":"afc14f30d7a66a863c473afd0f6615629602bb7edbb3434dd8a526a12fd24177"} Mar 20 09:38:04 crc kubenswrapper[4858]: I0320 09:38:04.721012 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566658-tht4s" Mar 20 09:38:04 crc kubenswrapper[4858]: I0320 09:38:04.721022 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afc14f30d7a66a863c473afd0f6615629602bb7edbb3434dd8a526a12fd24177" Mar 20 09:38:05 crc kubenswrapper[4858]: I0320 09:38:05.115978 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-nzq49"] Mar 20 09:38:05 crc kubenswrapper[4858]: I0320 09:38:05.128565 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566652-nzq49"] Mar 20 09:38:06 crc kubenswrapper[4858]: I0320 09:38:06.082078 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="773714cb-9045-4504-8813-9f389ed64883" path="/var/lib/kubelet/pods/773714cb-9045-4504-8813-9f389ed64883/volumes" Mar 20 09:38:07 crc kubenswrapper[4858]: I0320 09:38:07.890150 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:38:07 crc kubenswrapper[4858]: I0320 09:38:07.890672 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:38:31 crc kubenswrapper[4858]: I0320 09:38:31.197992 4858 scope.go:117] "RemoveContainer" containerID="2388726181deac0f3720191eea32ad01bbcb2388a8f20ae64a1684ce7835bd53" Mar 20 09:38:37 crc kubenswrapper[4858]: I0320 09:38:37.890627 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:38:37 crc kubenswrapper[4858]: I0320 09:38:37.891164 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:38:37 crc kubenswrapper[4858]: I0320 09:38:37.891217 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:38:37 crc kubenswrapper[4858]: I0320 09:38:37.891934 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:38:37 crc kubenswrapper[4858]: I0320 09:38:37.891990 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" gracePeriod=600 Mar 20 09:38:38 crc kubenswrapper[4858]: E0320 09:38:38.016110 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:38:38 crc kubenswrapper[4858]: I0320 09:38:38.988830 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" exitCode=0 Mar 20 09:38:38 crc kubenswrapper[4858]: I0320 09:38:38.988896 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6"} Mar 20 09:38:38 crc kubenswrapper[4858]: I0320 09:38:38.989290 4858 scope.go:117] "RemoveContainer" containerID="5817ed13dcb1fb797014d77cda07222984a083879566a3f0d5db8e9e973bd960" Mar 20 09:38:38 crc kubenswrapper[4858]: I0320 09:38:38.989711 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:38:38 crc kubenswrapper[4858]: E0320 09:38:38.990024 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:38:53 crc kubenswrapper[4858]: I0320 09:38:53.070830 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:38:53 crc kubenswrapper[4858]: E0320 09:38:53.071434 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:39:07 crc kubenswrapper[4858]: I0320 09:39:07.070501 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:39:07 crc kubenswrapper[4858]: E0320 09:39:07.071689 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:39:22 crc kubenswrapper[4858]: I0320 09:39:22.070857 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:39:22 crc kubenswrapper[4858]: E0320 09:39:22.073675 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:39:37 crc kubenswrapper[4858]: I0320 09:39:37.070564 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:39:37 crc kubenswrapper[4858]: E0320 09:39:37.071913 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:39:52 crc kubenswrapper[4858]: I0320 09:39:52.069776 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:39:52 crc kubenswrapper[4858]: E0320 09:39:52.070446 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.142577 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566660-hxxdt"] Mar 20 09:40:00 crc kubenswrapper[4858]: E0320 09:40:00.143363 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1daa8fc3-43fa-4b60-a94a-4594306737d5" containerName="oc" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.143378 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1daa8fc3-43fa-4b60-a94a-4594306737d5" containerName="oc" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.143534 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="1daa8fc3-43fa-4b60-a94a-4594306737d5" containerName="oc" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.144092 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-hxxdt" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.147536 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.147625 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.147691 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.186082 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-hxxdt"] Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.228300 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4vpq\" (UniqueName: \"kubernetes.io/projected/b75ada44-8bbf-47bd-a20b-eb9d95138980-kube-api-access-w4vpq\") pod \"auto-csr-approver-29566660-hxxdt\" (UID: \"b75ada44-8bbf-47bd-a20b-eb9d95138980\") " pod="openshift-infra/auto-csr-approver-29566660-hxxdt" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.330172 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4vpq\" (UniqueName: \"kubernetes.io/projected/b75ada44-8bbf-47bd-a20b-eb9d95138980-kube-api-access-w4vpq\") pod \"auto-csr-approver-29566660-hxxdt\" (UID: \"b75ada44-8bbf-47bd-a20b-eb9d95138980\") " pod="openshift-infra/auto-csr-approver-29566660-hxxdt" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.351563 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4vpq\" (UniqueName: \"kubernetes.io/projected/b75ada44-8bbf-47bd-a20b-eb9d95138980-kube-api-access-w4vpq\") pod \"auto-csr-approver-29566660-hxxdt\" (UID: \"b75ada44-8bbf-47bd-a20b-eb9d95138980\") " pod="openshift-infra/auto-csr-approver-29566660-hxxdt" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.466669 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-hxxdt" Mar 20 09:40:00 crc kubenswrapper[4858]: I0320 09:40:00.885989 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-hxxdt"] Mar 20 09:40:01 crc kubenswrapper[4858]: I0320 09:40:01.690467 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566660-hxxdt" event={"ID":"b75ada44-8bbf-47bd-a20b-eb9d95138980","Type":"ContainerStarted","Data":"8046fecccfbbbc417f666cc8880db0ed167b9f83e65e296585eccdf6c1fa98aa"} Mar 20 09:40:02 crc kubenswrapper[4858]: I0320 09:40:02.699101 4858 generic.go:334] "Generic (PLEG): container finished" podID="b75ada44-8bbf-47bd-a20b-eb9d95138980" containerID="714e78437801933d5a346d4ef966b0dff0682fd7acda463de7d70bec21dea9f5" exitCode=0 Mar 20 09:40:02 crc kubenswrapper[4858]: I0320 09:40:02.699373 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566660-hxxdt" event={"ID":"b75ada44-8bbf-47bd-a20b-eb9d95138980","Type":"ContainerDied","Data":"714e78437801933d5a346d4ef966b0dff0682fd7acda463de7d70bec21dea9f5"} Mar 20 09:40:03 crc kubenswrapper[4858]: I0320 09:40:03.070052 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:40:03 crc kubenswrapper[4858]: E0320 09:40:03.070363 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:40:04 crc kubenswrapper[4858]: I0320 09:40:04.035367 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-hxxdt" Mar 20 09:40:04 crc kubenswrapper[4858]: I0320 09:40:04.183753 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4vpq\" (UniqueName: \"kubernetes.io/projected/b75ada44-8bbf-47bd-a20b-eb9d95138980-kube-api-access-w4vpq\") pod \"b75ada44-8bbf-47bd-a20b-eb9d95138980\" (UID: \"b75ada44-8bbf-47bd-a20b-eb9d95138980\") " Mar 20 09:40:04 crc kubenswrapper[4858]: I0320 09:40:04.191404 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b75ada44-8bbf-47bd-a20b-eb9d95138980-kube-api-access-w4vpq" (OuterVolumeSpecName: "kube-api-access-w4vpq") pod "b75ada44-8bbf-47bd-a20b-eb9d95138980" (UID: "b75ada44-8bbf-47bd-a20b-eb9d95138980"). InnerVolumeSpecName "kube-api-access-w4vpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:40:04 crc kubenswrapper[4858]: I0320 09:40:04.285504 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4vpq\" (UniqueName: \"kubernetes.io/projected/b75ada44-8bbf-47bd-a20b-eb9d95138980-kube-api-access-w4vpq\") on node \"crc\" DevicePath \"\"" Mar 20 09:40:04 crc kubenswrapper[4858]: I0320 09:40:04.714256 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566660-hxxdt" event={"ID":"b75ada44-8bbf-47bd-a20b-eb9d95138980","Type":"ContainerDied","Data":"8046fecccfbbbc417f666cc8880db0ed167b9f83e65e296585eccdf6c1fa98aa"} Mar 20 09:40:04 crc kubenswrapper[4858]: I0320 09:40:04.714304 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8046fecccfbbbc417f666cc8880db0ed167b9f83e65e296585eccdf6c1fa98aa" Mar 20 09:40:04 crc kubenswrapper[4858]: I0320 09:40:04.714352 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566660-hxxdt" Mar 20 09:40:05 crc kubenswrapper[4858]: I0320 09:40:05.104067 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-6ht6p"] Mar 20 09:40:05 crc kubenswrapper[4858]: I0320 09:40:05.109834 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566654-6ht6p"] Mar 20 09:40:06 crc kubenswrapper[4858]: I0320 09:40:06.080372 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d98bc030-f3d4-4753-b42b-f7674adc1816" path="/var/lib/kubelet/pods/d98bc030-f3d4-4753-b42b-f7674adc1816/volumes" Mar 20 09:40:14 crc kubenswrapper[4858]: I0320 09:40:14.070703 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:40:14 crc kubenswrapper[4858]: E0320 09:40:14.071688 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:40:25 crc kubenswrapper[4858]: I0320 09:40:25.069988 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:40:25 crc kubenswrapper[4858]: E0320 09:40:25.070779 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:40:31 crc kubenswrapper[4858]: I0320 09:40:31.276634 4858 scope.go:117] "RemoveContainer" containerID="c0003c9bd93e1176ada49158eb4c75ce389393064d2ed2e7bfa7638cd85f4fde" Mar 20 09:40:36 crc kubenswrapper[4858]: I0320 09:40:36.070706 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:40:36 crc kubenswrapper[4858]: E0320 09:40:36.071565 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:40:49 crc kubenswrapper[4858]: I0320 09:40:49.070039 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:40:49 crc kubenswrapper[4858]: E0320 09:40:49.070941 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:41:04 crc kubenswrapper[4858]: I0320 09:41:04.069963 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:41:04 crc kubenswrapper[4858]: E0320 09:41:04.071250 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:41:18 crc kubenswrapper[4858]: I0320 09:41:18.070730 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:41:18 crc kubenswrapper[4858]: E0320 09:41:18.071710 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:41:30 crc kubenswrapper[4858]: I0320 09:41:30.075262 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:41:30 crc kubenswrapper[4858]: E0320 09:41:30.076162 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:41:45 crc kubenswrapper[4858]: I0320 09:41:45.069749 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:41:45 crc kubenswrapper[4858]: E0320 09:41:45.070590 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:41:58 crc kubenswrapper[4858]: I0320 09:41:58.070581 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:41:58 crc kubenswrapper[4858]: E0320 09:41:58.071657 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.151747 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566662-gx9st"] Mar 20 09:42:00 crc kubenswrapper[4858]: E0320 09:42:00.152448 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b75ada44-8bbf-47bd-a20b-eb9d95138980" containerName="oc" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.152466 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="b75ada44-8bbf-47bd-a20b-eb9d95138980" containerName="oc" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.152634 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="b75ada44-8bbf-47bd-a20b-eb9d95138980" containerName="oc" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.153088 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-gx9st" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.155465 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.155537 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.158226 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-gx9st"] Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.158829 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.315380 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbcjq\" (UniqueName: \"kubernetes.io/projected/cc444685-16df-4295-b62a-c0d9e26af1d7-kube-api-access-fbcjq\") pod \"auto-csr-approver-29566662-gx9st\" (UID: \"cc444685-16df-4295-b62a-c0d9e26af1d7\") " pod="openshift-infra/auto-csr-approver-29566662-gx9st" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.417190 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbcjq\" (UniqueName: \"kubernetes.io/projected/cc444685-16df-4295-b62a-c0d9e26af1d7-kube-api-access-fbcjq\") pod \"auto-csr-approver-29566662-gx9st\" (UID: \"cc444685-16df-4295-b62a-c0d9e26af1d7\") " pod="openshift-infra/auto-csr-approver-29566662-gx9st" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.442585 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbcjq\" (UniqueName: \"kubernetes.io/projected/cc444685-16df-4295-b62a-c0d9e26af1d7-kube-api-access-fbcjq\") pod \"auto-csr-approver-29566662-gx9st\" (UID: \"cc444685-16df-4295-b62a-c0d9e26af1d7\") " pod="openshift-infra/auto-csr-approver-29566662-gx9st" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.469440 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-gx9st" Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.902760 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-gx9st"] Mar 20 09:42:00 crc kubenswrapper[4858]: I0320 09:42:00.909724 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:42:01 crc kubenswrapper[4858]: I0320 09:42:01.622016 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566662-gx9st" event={"ID":"cc444685-16df-4295-b62a-c0d9e26af1d7","Type":"ContainerStarted","Data":"68e280daa95d00ecb4b0dae9d3abd1b0c4a5c14880ca31338fa6f715839a9a0a"} Mar 20 09:42:02 crc kubenswrapper[4858]: I0320 09:42:02.630939 4858 generic.go:334] "Generic (PLEG): container finished" podID="cc444685-16df-4295-b62a-c0d9e26af1d7" containerID="ebca739e8fd0b796211a78f1349bf65fe26e90053909bfe64d7549f0fd74e392" exitCode=0 Mar 20 09:42:02 crc kubenswrapper[4858]: I0320 09:42:02.631295 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566662-gx9st" event={"ID":"cc444685-16df-4295-b62a-c0d9e26af1d7","Type":"ContainerDied","Data":"ebca739e8fd0b796211a78f1349bf65fe26e90053909bfe64d7549f0fd74e392"} Mar 20 09:42:04 crc kubenswrapper[4858]: I0320 09:42:04.026264 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-gx9st" Mar 20 09:42:04 crc kubenswrapper[4858]: I0320 09:42:04.084954 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbcjq\" (UniqueName: \"kubernetes.io/projected/cc444685-16df-4295-b62a-c0d9e26af1d7-kube-api-access-fbcjq\") pod \"cc444685-16df-4295-b62a-c0d9e26af1d7\" (UID: \"cc444685-16df-4295-b62a-c0d9e26af1d7\") " Mar 20 09:42:04 crc kubenswrapper[4858]: I0320 09:42:04.091969 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc444685-16df-4295-b62a-c0d9e26af1d7-kube-api-access-fbcjq" (OuterVolumeSpecName: "kube-api-access-fbcjq") pod "cc444685-16df-4295-b62a-c0d9e26af1d7" (UID: "cc444685-16df-4295-b62a-c0d9e26af1d7"). InnerVolumeSpecName "kube-api-access-fbcjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:42:04 crc kubenswrapper[4858]: I0320 09:42:04.186654 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbcjq\" (UniqueName: \"kubernetes.io/projected/cc444685-16df-4295-b62a-c0d9e26af1d7-kube-api-access-fbcjq\") on node \"crc\" DevicePath \"\"" Mar 20 09:42:04 crc kubenswrapper[4858]: I0320 09:42:04.647099 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566662-gx9st" event={"ID":"cc444685-16df-4295-b62a-c0d9e26af1d7","Type":"ContainerDied","Data":"68e280daa95d00ecb4b0dae9d3abd1b0c4a5c14880ca31338fa6f715839a9a0a"} Mar 20 09:42:04 crc kubenswrapper[4858]: I0320 09:42:04.647535 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68e280daa95d00ecb4b0dae9d3abd1b0c4a5c14880ca31338fa6f715839a9a0a" Mar 20 09:42:04 crc kubenswrapper[4858]: I0320 09:42:04.647175 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566662-gx9st" Mar 20 09:42:05 crc kubenswrapper[4858]: I0320 09:42:05.099619 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-ntc4k"] Mar 20 09:42:05 crc kubenswrapper[4858]: I0320 09:42:05.106443 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566656-ntc4k"] Mar 20 09:42:06 crc kubenswrapper[4858]: I0320 09:42:06.077583 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ff6909-89ba-4b73-9940-eb06289f2a30" path="/var/lib/kubelet/pods/08ff6909-89ba-4b73-9940-eb06289f2a30/volumes" Mar 20 09:42:09 crc kubenswrapper[4858]: I0320 09:42:09.071875 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:42:09 crc kubenswrapper[4858]: E0320 09:42:09.072568 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:42:20 crc kubenswrapper[4858]: I0320 09:42:20.075087 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:42:20 crc kubenswrapper[4858]: E0320 09:42:20.076217 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:42:31 crc kubenswrapper[4858]: I0320 09:42:31.374120 4858 scope.go:117] "RemoveContainer" containerID="101c55c930a879083435516e891376919f72cbb9e76b90005e15b16685d68814" Mar 20 09:42:35 crc kubenswrapper[4858]: I0320 09:42:35.071145 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:42:35 crc kubenswrapper[4858]: E0320 09:42:35.072226 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:42:50 crc kubenswrapper[4858]: I0320 09:42:50.075581 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:42:50 crc kubenswrapper[4858]: E0320 09:42:50.076656 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:43:04 crc kubenswrapper[4858]: I0320 09:43:04.070856 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:43:04 crc kubenswrapper[4858]: E0320 09:43:04.072083 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:43:18 crc kubenswrapper[4858]: I0320 09:43:18.070685 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:43:18 crc kubenswrapper[4858]: E0320 09:43:18.072138 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:43:31 crc kubenswrapper[4858]: I0320 09:43:31.070393 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:43:31 crc kubenswrapper[4858]: E0320 09:43:31.071115 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:43:44 crc kubenswrapper[4858]: I0320 09:43:44.070650 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:43:44 crc kubenswrapper[4858]: I0320 09:43:44.445633 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"021bb0318f7c2654b5dd7a30ce0569091d573b7357c04b6dbe28f1b4c1f439c8"} Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.155589 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566664-g4tn4"] Mar 20 09:44:00 crc kubenswrapper[4858]: E0320 09:44:00.156793 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc444685-16df-4295-b62a-c0d9e26af1d7" containerName="oc" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.156819 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc444685-16df-4295-b62a-c0d9e26af1d7" containerName="oc" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.157079 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc444685-16df-4295-b62a-c0d9e26af1d7" containerName="oc" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.157938 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-g4tn4" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.162689 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-g4tn4"] Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.162892 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.163273 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.163670 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.313979 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqpwb\" (UniqueName: \"kubernetes.io/projected/065a313a-b2c9-4b27-b063-bd71ed126811-kube-api-access-jqpwb\") pod \"auto-csr-approver-29566664-g4tn4\" (UID: \"065a313a-b2c9-4b27-b063-bd71ed126811\") " pod="openshift-infra/auto-csr-approver-29566664-g4tn4" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.415786 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqpwb\" (UniqueName: \"kubernetes.io/projected/065a313a-b2c9-4b27-b063-bd71ed126811-kube-api-access-jqpwb\") pod \"auto-csr-approver-29566664-g4tn4\" (UID: \"065a313a-b2c9-4b27-b063-bd71ed126811\") " pod="openshift-infra/auto-csr-approver-29566664-g4tn4" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.437190 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqpwb\" (UniqueName: \"kubernetes.io/projected/065a313a-b2c9-4b27-b063-bd71ed126811-kube-api-access-jqpwb\") pod \"auto-csr-approver-29566664-g4tn4\" (UID: \"065a313a-b2c9-4b27-b063-bd71ed126811\") " pod="openshift-infra/auto-csr-approver-29566664-g4tn4" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.482725 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-g4tn4" Mar 20 09:44:00 crc kubenswrapper[4858]: I0320 09:44:00.912332 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-g4tn4"] Mar 20 09:44:01 crc kubenswrapper[4858]: I0320 09:44:01.600118 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566664-g4tn4" event={"ID":"065a313a-b2c9-4b27-b063-bd71ed126811","Type":"ContainerStarted","Data":"e5cc2dc7aa1c445b5653f791ec9f26ca5c40cff890ed173f953e1918cbd4b1f4"} Mar 20 09:44:02 crc kubenswrapper[4858]: I0320 09:44:02.608541 4858 generic.go:334] "Generic (PLEG): container finished" podID="065a313a-b2c9-4b27-b063-bd71ed126811" containerID="47cba87d5c4976d6c4bf65ca3efe1bc5c981ce059cea321eb3a1c12d6ba8d8c3" exitCode=0 Mar 20 09:44:02 crc kubenswrapper[4858]: I0320 09:44:02.608619 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566664-g4tn4" event={"ID":"065a313a-b2c9-4b27-b063-bd71ed126811","Type":"ContainerDied","Data":"47cba87d5c4976d6c4bf65ca3efe1bc5c981ce059cea321eb3a1c12d6ba8d8c3"} Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.013039 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-g4tn4" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.063513 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqpwb\" (UniqueName: \"kubernetes.io/projected/065a313a-b2c9-4b27-b063-bd71ed126811-kube-api-access-jqpwb\") pod \"065a313a-b2c9-4b27-b063-bd71ed126811\" (UID: \"065a313a-b2c9-4b27-b063-bd71ed126811\") " Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.074108 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/065a313a-b2c9-4b27-b063-bd71ed126811-kube-api-access-jqpwb" (OuterVolumeSpecName: "kube-api-access-jqpwb") pod "065a313a-b2c9-4b27-b063-bd71ed126811" (UID: "065a313a-b2c9-4b27-b063-bd71ed126811"). InnerVolumeSpecName "kube-api-access-jqpwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.165100 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqpwb\" (UniqueName: \"kubernetes.io/projected/065a313a-b2c9-4b27-b063-bd71ed126811-kube-api-access-jqpwb\") on node \"crc\" DevicePath \"\"" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.513464 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dbs7t"] Mar 20 09:44:04 crc kubenswrapper[4858]: E0320 09:44:04.513944 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="065a313a-b2c9-4b27-b063-bd71ed126811" containerName="oc" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.513991 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="065a313a-b2c9-4b27-b063-bd71ed126811" containerName="oc" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.514209 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="065a313a-b2c9-4b27-b063-bd71ed126811" containerName="oc" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.515445 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.528763 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dbs7t"] Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.572128 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-utilities\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.572174 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7drrt\" (UniqueName: \"kubernetes.io/projected/2f90972c-8d0d-46ec-82dd-ab9157cc6436-kube-api-access-7drrt\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.572201 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-catalog-content\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.622287 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566664-g4tn4" event={"ID":"065a313a-b2c9-4b27-b063-bd71ed126811","Type":"ContainerDied","Data":"e5cc2dc7aa1c445b5653f791ec9f26ca5c40cff890ed173f953e1918cbd4b1f4"} Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.622343 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5cc2dc7aa1c445b5653f791ec9f26ca5c40cff890ed173f953e1918cbd4b1f4" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.622402 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566664-g4tn4" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.673937 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-utilities\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.673989 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7drrt\" (UniqueName: \"kubernetes.io/projected/2f90972c-8d0d-46ec-82dd-ab9157cc6436-kube-api-access-7drrt\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.674024 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-catalog-content\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.674531 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-utilities\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.674596 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-catalog-content\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.692174 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7drrt\" (UniqueName: \"kubernetes.io/projected/2f90972c-8d0d-46ec-82dd-ab9157cc6436-kube-api-access-7drrt\") pod \"redhat-operators-dbs7t\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:04 crc kubenswrapper[4858]: I0320 09:44:04.834026 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:05 crc kubenswrapper[4858]: I0320 09:44:05.085484 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-tht4s"] Mar 20 09:44:05 crc kubenswrapper[4858]: I0320 09:44:05.092254 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566658-tht4s"] Mar 20 09:44:05 crc kubenswrapper[4858]: I0320 09:44:05.268964 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dbs7t"] Mar 20 09:44:05 crc kubenswrapper[4858]: I0320 09:44:05.630663 4858 generic.go:334] "Generic (PLEG): container finished" podID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerID="bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3" exitCode=0 Mar 20 09:44:05 crc kubenswrapper[4858]: I0320 09:44:05.630723 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbs7t" event={"ID":"2f90972c-8d0d-46ec-82dd-ab9157cc6436","Type":"ContainerDied","Data":"bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3"} Mar 20 09:44:05 crc kubenswrapper[4858]: I0320 09:44:05.630971 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbs7t" event={"ID":"2f90972c-8d0d-46ec-82dd-ab9157cc6436","Type":"ContainerStarted","Data":"9c9ffe38d99a4fd45f81a7878113292b5694e73ba56fef8f45e85ade2a002573"} Mar 20 09:44:06 crc kubenswrapper[4858]: I0320 09:44:06.078752 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1daa8fc3-43fa-4b60-a94a-4594306737d5" path="/var/lib/kubelet/pods/1daa8fc3-43fa-4b60-a94a-4594306737d5/volumes" Mar 20 09:44:06 crc kubenswrapper[4858]: I0320 09:44:06.640093 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbs7t" event={"ID":"2f90972c-8d0d-46ec-82dd-ab9157cc6436","Type":"ContainerStarted","Data":"56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9"} Mar 20 09:44:07 crc kubenswrapper[4858]: I0320 09:44:07.651846 4858 generic.go:334] "Generic (PLEG): container finished" podID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerID="56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9" exitCode=0 Mar 20 09:44:07 crc kubenswrapper[4858]: I0320 09:44:07.651898 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbs7t" event={"ID":"2f90972c-8d0d-46ec-82dd-ab9157cc6436","Type":"ContainerDied","Data":"56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9"} Mar 20 09:44:08 crc kubenswrapper[4858]: I0320 09:44:08.661703 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbs7t" event={"ID":"2f90972c-8d0d-46ec-82dd-ab9157cc6436","Type":"ContainerStarted","Data":"23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79"} Mar 20 09:44:08 crc kubenswrapper[4858]: I0320 09:44:08.682061 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dbs7t" podStartSLOduration=2.249496071 podStartE2EDuration="4.682041827s" podCreationTimestamp="2026-03-20 09:44:04 +0000 UTC" firstStartedPulling="2026-03-20 09:44:05.632206642 +0000 UTC m=+2826.952624839" lastFinishedPulling="2026-03-20 09:44:08.064752398 +0000 UTC m=+2829.385170595" observedRunningTime="2026-03-20 09:44:08.678417933 +0000 UTC m=+2829.998836150" watchObservedRunningTime="2026-03-20 09:44:08.682041827 +0000 UTC m=+2830.002460024" Mar 20 09:44:14 crc kubenswrapper[4858]: I0320 09:44:14.835068 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:14 crc kubenswrapper[4858]: I0320 09:44:14.835670 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:15 crc kubenswrapper[4858]: I0320 09:44:15.887479 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dbs7t" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="registry-server" probeResult="failure" output=< Mar 20 09:44:15 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:44:15 crc kubenswrapper[4858]: > Mar 20 09:44:24 crc kubenswrapper[4858]: I0320 09:44:24.884201 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:24 crc kubenswrapper[4858]: I0320 09:44:24.926127 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.188220 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dbs7t"] Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.189034 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dbs7t" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="registry-server" containerID="cri-o://23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79" gracePeriod=2 Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.723837 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.804203 4858 generic.go:334] "Generic (PLEG): container finished" podID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerID="23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79" exitCode=0 Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.804245 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbs7t" event={"ID":"2f90972c-8d0d-46ec-82dd-ab9157cc6436","Type":"ContainerDied","Data":"23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79"} Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.804261 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dbs7t" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.804281 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dbs7t" event={"ID":"2f90972c-8d0d-46ec-82dd-ab9157cc6436","Type":"ContainerDied","Data":"9c9ffe38d99a4fd45f81a7878113292b5694e73ba56fef8f45e85ade2a002573"} Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.804302 4858 scope.go:117] "RemoveContainer" containerID="23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.823986 4858 scope.go:117] "RemoveContainer" containerID="56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.843390 4858 scope.go:117] "RemoveContainer" containerID="bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.867904 4858 scope.go:117] "RemoveContainer" containerID="23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79" Mar 20 09:44:27 crc kubenswrapper[4858]: E0320 09:44:27.868471 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79\": container with ID starting with 23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79 not found: ID does not exist" containerID="23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.868524 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79"} err="failed to get container status \"23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79\": rpc error: code = NotFound desc = could not find container \"23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79\": container with ID starting with 23928f0d8300a5223a242f2f809425ea8dfdeb943feb8c802f234c9864b28e79 not found: ID does not exist" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.868557 4858 scope.go:117] "RemoveContainer" containerID="56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9" Mar 20 09:44:27 crc kubenswrapper[4858]: E0320 09:44:27.869033 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9\": container with ID starting with 56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9 not found: ID does not exist" containerID="56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.869070 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9"} err="failed to get container status \"56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9\": rpc error: code = NotFound desc = could not find container \"56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9\": container with ID starting with 56a2f89e0228141efa1ded8aa3f5be76696e0eea1b16bbf1d4fbfb2e62adedd9 not found: ID does not exist" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.869095 4858 scope.go:117] "RemoveContainer" containerID="bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3" Mar 20 09:44:27 crc kubenswrapper[4858]: E0320 09:44:27.869647 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3\": container with ID starting with bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3 not found: ID does not exist" containerID="bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.869678 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3"} err="failed to get container status \"bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3\": rpc error: code = NotFound desc = could not find container \"bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3\": container with ID starting with bb822bb435365ed36ed737968357256a3273452214dea679cb86ed77136a47b3 not found: ID does not exist" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.915024 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7drrt\" (UniqueName: \"kubernetes.io/projected/2f90972c-8d0d-46ec-82dd-ab9157cc6436-kube-api-access-7drrt\") pod \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.915081 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-utilities\") pod \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.915196 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-catalog-content\") pod \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\" (UID: \"2f90972c-8d0d-46ec-82dd-ab9157cc6436\") " Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.916357 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-utilities" (OuterVolumeSpecName: "utilities") pod "2f90972c-8d0d-46ec-82dd-ab9157cc6436" (UID: "2f90972c-8d0d-46ec-82dd-ab9157cc6436"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:44:27 crc kubenswrapper[4858]: I0320 09:44:27.921024 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f90972c-8d0d-46ec-82dd-ab9157cc6436-kube-api-access-7drrt" (OuterVolumeSpecName: "kube-api-access-7drrt") pod "2f90972c-8d0d-46ec-82dd-ab9157cc6436" (UID: "2f90972c-8d0d-46ec-82dd-ab9157cc6436"). InnerVolumeSpecName "kube-api-access-7drrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:44:28 crc kubenswrapper[4858]: I0320 09:44:28.016944 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7drrt\" (UniqueName: \"kubernetes.io/projected/2f90972c-8d0d-46ec-82dd-ab9157cc6436-kube-api-access-7drrt\") on node \"crc\" DevicePath \"\"" Mar 20 09:44:28 crc kubenswrapper[4858]: I0320 09:44:28.017008 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:44:28 crc kubenswrapper[4858]: I0320 09:44:28.056299 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2f90972c-8d0d-46ec-82dd-ab9157cc6436" (UID: "2f90972c-8d0d-46ec-82dd-ab9157cc6436"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:44:28 crc kubenswrapper[4858]: I0320 09:44:28.123538 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2f90972c-8d0d-46ec-82dd-ab9157cc6436-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:44:28 crc kubenswrapper[4858]: I0320 09:44:28.128408 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dbs7t"] Mar 20 09:44:28 crc kubenswrapper[4858]: I0320 09:44:28.133809 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dbs7t"] Mar 20 09:44:30 crc kubenswrapper[4858]: I0320 09:44:30.080219 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" path="/var/lib/kubelet/pods/2f90972c-8d0d-46ec-82dd-ab9157cc6436/volumes" Mar 20 09:44:31 crc kubenswrapper[4858]: I0320 09:44:31.457861 4858 scope.go:117] "RemoveContainer" containerID="62fa5939ee5b6f50beb2399d7ce0432f46efdffdda67784869b7fb84502a48fd" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.155529 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh"] Mar 20 09:45:00 crc kubenswrapper[4858]: E0320 09:45:00.158251 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="extract-content" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.158410 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="extract-content" Mar 20 09:45:00 crc kubenswrapper[4858]: E0320 09:45:00.158534 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="extract-utilities" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.158625 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="extract-utilities" Mar 20 09:45:00 crc kubenswrapper[4858]: E0320 09:45:00.158708 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="registry-server" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.158818 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="registry-server" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.159122 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f90972c-8d0d-46ec-82dd-ab9157cc6436" containerName="registry-server" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.159844 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.163582 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.163628 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.165394 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh"] Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.244131 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgbr6\" (UniqueName: \"kubernetes.io/projected/2c833074-ddaa-445a-8c05-b6c4d02a51b9-kube-api-access-rgbr6\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.244243 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c833074-ddaa-445a-8c05-b6c4d02a51b9-secret-volume\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.244337 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c833074-ddaa-445a-8c05-b6c4d02a51b9-config-volume\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.345587 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgbr6\" (UniqueName: \"kubernetes.io/projected/2c833074-ddaa-445a-8c05-b6c4d02a51b9-kube-api-access-rgbr6\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.345704 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c833074-ddaa-445a-8c05-b6c4d02a51b9-secret-volume\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.345783 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c833074-ddaa-445a-8c05-b6c4d02a51b9-config-volume\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.347916 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c833074-ddaa-445a-8c05-b6c4d02a51b9-config-volume\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.354383 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c833074-ddaa-445a-8c05-b6c4d02a51b9-secret-volume\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.367877 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgbr6\" (UniqueName: \"kubernetes.io/projected/2c833074-ddaa-445a-8c05-b6c4d02a51b9-kube-api-access-rgbr6\") pod \"collect-profiles-29566665-wrmxh\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.491989 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:00 crc kubenswrapper[4858]: I0320 09:45:00.946868 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh"] Mar 20 09:45:01 crc kubenswrapper[4858]: I0320 09:45:01.131302 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" event={"ID":"2c833074-ddaa-445a-8c05-b6c4d02a51b9","Type":"ContainerStarted","Data":"2a3df5570e1856612e0e7a2b9806010146b329e34806f8af53d2491143e6cfa9"} Mar 20 09:45:02 crc kubenswrapper[4858]: I0320 09:45:02.146220 4858 generic.go:334] "Generic (PLEG): container finished" podID="2c833074-ddaa-445a-8c05-b6c4d02a51b9" containerID="7a0cfd3d08daf22d6a37d78f67e305b35d5baab529a0821dc02fee882b5b4aea" exitCode=0 Mar 20 09:45:02 crc kubenswrapper[4858]: I0320 09:45:02.147843 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" event={"ID":"2c833074-ddaa-445a-8c05-b6c4d02a51b9","Type":"ContainerDied","Data":"7a0cfd3d08daf22d6a37d78f67e305b35d5baab529a0821dc02fee882b5b4aea"} Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.556192 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.618605 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c833074-ddaa-445a-8c05-b6c4d02a51b9-config-volume\") pod \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.619758 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c833074-ddaa-445a-8c05-b6c4d02a51b9-config-volume" (OuterVolumeSpecName: "config-volume") pod "2c833074-ddaa-445a-8c05-b6c4d02a51b9" (UID: "2c833074-ddaa-445a-8c05-b6c4d02a51b9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.719714 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c833074-ddaa-445a-8c05-b6c4d02a51b9-secret-volume\") pod \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.719827 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgbr6\" (UniqueName: \"kubernetes.io/projected/2c833074-ddaa-445a-8c05-b6c4d02a51b9-kube-api-access-rgbr6\") pod \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\" (UID: \"2c833074-ddaa-445a-8c05-b6c4d02a51b9\") " Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.720211 4858 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2c833074-ddaa-445a-8c05-b6c4d02a51b9-config-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.729242 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c833074-ddaa-445a-8c05-b6c4d02a51b9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2c833074-ddaa-445a-8c05-b6c4d02a51b9" (UID: "2c833074-ddaa-445a-8c05-b6c4d02a51b9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.730673 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c833074-ddaa-445a-8c05-b6c4d02a51b9-kube-api-access-rgbr6" (OuterVolumeSpecName: "kube-api-access-rgbr6") pod "2c833074-ddaa-445a-8c05-b6c4d02a51b9" (UID: "2c833074-ddaa-445a-8c05-b6c4d02a51b9"). InnerVolumeSpecName "kube-api-access-rgbr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.821062 4858 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2c833074-ddaa-445a-8c05-b6c4d02a51b9-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:03 crc kubenswrapper[4858]: I0320 09:45:03.821124 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgbr6\" (UniqueName: \"kubernetes.io/projected/2c833074-ddaa-445a-8c05-b6c4d02a51b9-kube-api-access-rgbr6\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:04 crc kubenswrapper[4858]: I0320 09:45:04.165887 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" event={"ID":"2c833074-ddaa-445a-8c05-b6c4d02a51b9","Type":"ContainerDied","Data":"2a3df5570e1856612e0e7a2b9806010146b329e34806f8af53d2491143e6cfa9"} Mar 20 09:45:04 crc kubenswrapper[4858]: I0320 09:45:04.165947 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a3df5570e1856612e0e7a2b9806010146b329e34806f8af53d2491143e6cfa9" Mar 20 09:45:04 crc kubenswrapper[4858]: I0320 09:45:04.166024 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29566665-wrmxh" Mar 20 09:45:04 crc kubenswrapper[4858]: I0320 09:45:04.695274 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd"] Mar 20 09:45:04 crc kubenswrapper[4858]: I0320 09:45:04.700337 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29566620-tjqsd"] Mar 20 09:45:06 crc kubenswrapper[4858]: I0320 09:45:06.083969 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8f37a14-0144-4154-a087-126fde1633eb" path="/var/lib/kubelet/pods/d8f37a14-0144-4154-a087-126fde1633eb/volumes" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.213023 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z4zr5"] Mar 20 09:45:12 crc kubenswrapper[4858]: E0320 09:45:12.214425 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c833074-ddaa-445a-8c05-b6c4d02a51b9" containerName="collect-profiles" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.214458 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c833074-ddaa-445a-8c05-b6c4d02a51b9" containerName="collect-profiles" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.214811 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c833074-ddaa-445a-8c05-b6c4d02a51b9" containerName="collect-profiles" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.216601 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.234451 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z4zr5"] Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.276973 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n54kt\" (UniqueName: \"kubernetes.io/projected/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-kube-api-access-n54kt\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.277065 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-catalog-content\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.277145 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-utilities\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.379264 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n54kt\" (UniqueName: \"kubernetes.io/projected/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-kube-api-access-n54kt\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.379390 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-catalog-content\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.379472 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-utilities\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.380005 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-utilities\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.380305 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-catalog-content\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.416052 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n54kt\" (UniqueName: \"kubernetes.io/projected/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-kube-api-access-n54kt\") pod \"community-operators-z4zr5\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:12 crc kubenswrapper[4858]: I0320 09:45:12.553096 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:13 crc kubenswrapper[4858]: I0320 09:45:13.100851 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z4zr5"] Mar 20 09:45:13 crc kubenswrapper[4858]: I0320 09:45:13.243661 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerStarted","Data":"1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a"} Mar 20 09:45:13 crc kubenswrapper[4858]: I0320 09:45:13.243913 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerStarted","Data":"c0fc2db1e260173241467958c205192e845e30795fceade80a4dbc9c0717d046"} Mar 20 09:45:14 crc kubenswrapper[4858]: I0320 09:45:14.254795 4858 generic.go:334] "Generic (PLEG): container finished" podID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerID="1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a" exitCode=0 Mar 20 09:45:14 crc kubenswrapper[4858]: I0320 09:45:14.254858 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerDied","Data":"1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a"} Mar 20 09:45:15 crc kubenswrapper[4858]: I0320 09:45:15.264740 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerStarted","Data":"1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648"} Mar 20 09:45:16 crc kubenswrapper[4858]: I0320 09:45:16.276292 4858 generic.go:334] "Generic (PLEG): container finished" podID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerID="1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648" exitCode=0 Mar 20 09:45:16 crc kubenswrapper[4858]: I0320 09:45:16.276369 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerDied","Data":"1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648"} Mar 20 09:45:17 crc kubenswrapper[4858]: I0320 09:45:17.286167 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerStarted","Data":"8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd"} Mar 20 09:45:17 crc kubenswrapper[4858]: I0320 09:45:17.316975 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z4zr5" podStartSLOduration=2.886091925 podStartE2EDuration="5.316956663s" podCreationTimestamp="2026-03-20 09:45:12 +0000 UTC" firstStartedPulling="2026-03-20 09:45:14.257667503 +0000 UTC m=+2895.578085730" lastFinishedPulling="2026-03-20 09:45:16.688532261 +0000 UTC m=+2898.008950468" observedRunningTime="2026-03-20 09:45:17.315842763 +0000 UTC m=+2898.636260970" watchObservedRunningTime="2026-03-20 09:45:17.316956663 +0000 UTC m=+2898.637374870" Mar 20 09:45:22 crc kubenswrapper[4858]: I0320 09:45:22.553570 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:22 crc kubenswrapper[4858]: I0320 09:45:22.554208 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:22 crc kubenswrapper[4858]: I0320 09:45:22.608269 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:23 crc kubenswrapper[4858]: I0320 09:45:23.406968 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:23 crc kubenswrapper[4858]: I0320 09:45:23.465503 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z4zr5"] Mar 20 09:45:25 crc kubenswrapper[4858]: I0320 09:45:25.352306 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z4zr5" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="registry-server" containerID="cri-o://8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd" gracePeriod=2 Mar 20 09:45:25 crc kubenswrapper[4858]: I0320 09:45:25.828406 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:25 crc kubenswrapper[4858]: I0320 09:45:25.997991 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n54kt\" (UniqueName: \"kubernetes.io/projected/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-kube-api-access-n54kt\") pod \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " Mar 20 09:45:25 crc kubenswrapper[4858]: I0320 09:45:25.998065 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-utilities\") pod \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " Mar 20 09:45:25 crc kubenswrapper[4858]: I0320 09:45:25.998242 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-catalog-content\") pod \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\" (UID: \"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e\") " Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.000702 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-utilities" (OuterVolumeSpecName: "utilities") pod "6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" (UID: "6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.008788 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-kube-api-access-n54kt" (OuterVolumeSpecName: "kube-api-access-n54kt") pod "6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" (UID: "6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e"). InnerVolumeSpecName "kube-api-access-n54kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.092252 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" (UID: "6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.103162 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.103196 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n54kt\" (UniqueName: \"kubernetes.io/projected/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-kube-api-access-n54kt\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.103212 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.364889 4858 generic.go:334] "Generic (PLEG): container finished" podID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerID="8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd" exitCode=0 Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.364968 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerDied","Data":"8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd"} Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.365724 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4zr5" event={"ID":"6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e","Type":"ContainerDied","Data":"c0fc2db1e260173241467958c205192e845e30795fceade80a4dbc9c0717d046"} Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.365009 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4zr5" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.365771 4858 scope.go:117] "RemoveContainer" containerID="8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.393651 4858 scope.go:117] "RemoveContainer" containerID="1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.420224 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z4zr5"] Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.427947 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z4zr5"] Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.435576 4858 scope.go:117] "RemoveContainer" containerID="1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.459776 4858 scope.go:117] "RemoveContainer" containerID="8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd" Mar 20 09:45:26 crc kubenswrapper[4858]: E0320 09:45:26.460522 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd\": container with ID starting with 8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd not found: ID does not exist" containerID="8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.460577 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd"} err="failed to get container status \"8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd\": rpc error: code = NotFound desc = could not find container \"8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd\": container with ID starting with 8f90f49a8b32b9b14461f4ff26b311d2f6f91600d2b4f881fd9d473a057044cd not found: ID does not exist" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.460626 4858 scope.go:117] "RemoveContainer" containerID="1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648" Mar 20 09:45:26 crc kubenswrapper[4858]: E0320 09:45:26.461333 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648\": container with ID starting with 1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648 not found: ID does not exist" containerID="1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.461388 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648"} err="failed to get container status \"1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648\": rpc error: code = NotFound desc = could not find container \"1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648\": container with ID starting with 1ea85966e74506b7e9a62e3401156ba920a2d996ae3dc44ffbd6328779783648 not found: ID does not exist" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.461418 4858 scope.go:117] "RemoveContainer" containerID="1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a" Mar 20 09:45:26 crc kubenswrapper[4858]: E0320 09:45:26.461818 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a\": container with ID starting with 1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a not found: ID does not exist" containerID="1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.461845 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a"} err="failed to get container status \"1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a\": rpc error: code = NotFound desc = could not find container \"1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a\": container with ID starting with 1fe3a9ab684afb00a3c8e1b4298a62be837f72c7955537b7cc9841f4d92a2f5a not found: ID does not exist" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.862378 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ns6qw"] Mar 20 09:45:26 crc kubenswrapper[4858]: E0320 09:45:26.862731 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="extract-content" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.862748 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="extract-content" Mar 20 09:45:26 crc kubenswrapper[4858]: E0320 09:45:26.862773 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="extract-utilities" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.862783 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="extract-utilities" Mar 20 09:45:26 crc kubenswrapper[4858]: E0320 09:45:26.862805 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="registry-server" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.862816 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="registry-server" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.863001 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" containerName="registry-server" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.864189 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.877308 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ns6qw"] Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.921102 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-utilities\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.921165 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-catalog-content\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:26 crc kubenswrapper[4858]: I0320 09:45:26.921197 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxn8n\" (UniqueName: \"kubernetes.io/projected/e6c57d77-ac27-46e2-958d-a1598a5de972-kube-api-access-kxn8n\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.023306 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-utilities\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.023668 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-catalog-content\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.023702 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxn8n\" (UniqueName: \"kubernetes.io/projected/e6c57d77-ac27-46e2-958d-a1598a5de972-kube-api-access-kxn8n\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.024061 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-utilities\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.024359 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-catalog-content\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.052610 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxn8n\" (UniqueName: \"kubernetes.io/projected/e6c57d77-ac27-46e2-958d-a1598a5de972-kube-api-access-kxn8n\") pod \"certified-operators-ns6qw\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.237565 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:27 crc kubenswrapper[4858]: I0320 09:45:27.663385 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ns6qw"] Mar 20 09:45:27 crc kubenswrapper[4858]: W0320 09:45:27.672518 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6c57d77_ac27_46e2_958d_a1598a5de972.slice/crio-802f07b473bd905d0a8e65178e053b02b9493aa4366a5aa15f4dcf7eed36f166 WatchSource:0}: Error finding container 802f07b473bd905d0a8e65178e053b02b9493aa4366a5aa15f4dcf7eed36f166: Status 404 returned error can't find the container with id 802f07b473bd905d0a8e65178e053b02b9493aa4366a5aa15f4dcf7eed36f166 Mar 20 09:45:28 crc kubenswrapper[4858]: I0320 09:45:28.085039 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e" path="/var/lib/kubelet/pods/6eec5f36-19e5-46b5-bc0e-dd612ee9cd0e/volumes" Mar 20 09:45:28 crc kubenswrapper[4858]: I0320 09:45:28.388909 4858 generic.go:334] "Generic (PLEG): container finished" podID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerID="302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb" exitCode=0 Mar 20 09:45:28 crc kubenswrapper[4858]: I0320 09:45:28.389115 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ns6qw" event={"ID":"e6c57d77-ac27-46e2-958d-a1598a5de972","Type":"ContainerDied","Data":"302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb"} Mar 20 09:45:28 crc kubenswrapper[4858]: I0320 09:45:28.389679 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ns6qw" event={"ID":"e6c57d77-ac27-46e2-958d-a1598a5de972","Type":"ContainerStarted","Data":"802f07b473bd905d0a8e65178e053b02b9493aa4366a5aa15f4dcf7eed36f166"} Mar 20 09:45:30 crc kubenswrapper[4858]: I0320 09:45:30.407368 4858 generic.go:334] "Generic (PLEG): container finished" podID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerID="0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc" exitCode=0 Mar 20 09:45:30 crc kubenswrapper[4858]: I0320 09:45:30.407494 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ns6qw" event={"ID":"e6c57d77-ac27-46e2-958d-a1598a5de972","Type":"ContainerDied","Data":"0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc"} Mar 20 09:45:31 crc kubenswrapper[4858]: I0320 09:45:31.423032 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ns6qw" event={"ID":"e6c57d77-ac27-46e2-958d-a1598a5de972","Type":"ContainerStarted","Data":"4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2"} Mar 20 09:45:31 crc kubenswrapper[4858]: I0320 09:45:31.446972 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ns6qw" podStartSLOduration=3.005176472 podStartE2EDuration="5.446936645s" podCreationTimestamp="2026-03-20 09:45:26 +0000 UTC" firstStartedPulling="2026-03-20 09:45:28.392387574 +0000 UTC m=+2909.712805801" lastFinishedPulling="2026-03-20 09:45:30.834147747 +0000 UTC m=+2912.154565974" observedRunningTime="2026-03-20 09:45:31.44637918 +0000 UTC m=+2912.766797377" watchObservedRunningTime="2026-03-20 09:45:31.446936645 +0000 UTC m=+2912.767354882" Mar 20 09:45:31 crc kubenswrapper[4858]: I0320 09:45:31.550665 4858 scope.go:117] "RemoveContainer" containerID="fe924a2f8c500c5a40f91694fab8c4b75b9ee5cad87fd09b14576ce594d745c0" Mar 20 09:45:37 crc kubenswrapper[4858]: I0320 09:45:37.238551 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:37 crc kubenswrapper[4858]: I0320 09:45:37.239238 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:37 crc kubenswrapper[4858]: I0320 09:45:37.298784 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:37 crc kubenswrapper[4858]: I0320 09:45:37.531141 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:38 crc kubenswrapper[4858]: I0320 09:45:38.032946 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ns6qw"] Mar 20 09:45:39 crc kubenswrapper[4858]: I0320 09:45:39.804580 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ns6qw" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="registry-server" containerID="cri-o://4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2" gracePeriod=2 Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.444789 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.541087 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-catalog-content\") pod \"e6c57d77-ac27-46e2-958d-a1598a5de972\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.541164 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-utilities\") pod \"e6c57d77-ac27-46e2-958d-a1598a5de972\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.541565 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxn8n\" (UniqueName: \"kubernetes.io/projected/e6c57d77-ac27-46e2-958d-a1598a5de972-kube-api-access-kxn8n\") pod \"e6c57d77-ac27-46e2-958d-a1598a5de972\" (UID: \"e6c57d77-ac27-46e2-958d-a1598a5de972\") " Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.542803 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-utilities" (OuterVolumeSpecName: "utilities") pod "e6c57d77-ac27-46e2-958d-a1598a5de972" (UID: "e6c57d77-ac27-46e2-958d-a1598a5de972"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.551642 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6c57d77-ac27-46e2-958d-a1598a5de972-kube-api-access-kxn8n" (OuterVolumeSpecName: "kube-api-access-kxn8n") pod "e6c57d77-ac27-46e2-958d-a1598a5de972" (UID: "e6c57d77-ac27-46e2-958d-a1598a5de972"). InnerVolumeSpecName "kube-api-access-kxn8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.594910 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6c57d77-ac27-46e2-958d-a1598a5de972" (UID: "e6c57d77-ac27-46e2-958d-a1598a5de972"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.644575 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxn8n\" (UniqueName: \"kubernetes.io/projected/e6c57d77-ac27-46e2-958d-a1598a5de972-kube-api-access-kxn8n\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.644621 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.644633 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6c57d77-ac27-46e2-958d-a1598a5de972-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.813002 4858 generic.go:334] "Generic (PLEG): container finished" podID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerID="4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2" exitCode=0 Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.813043 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ns6qw" event={"ID":"e6c57d77-ac27-46e2-958d-a1598a5de972","Type":"ContainerDied","Data":"4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2"} Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.813068 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ns6qw" event={"ID":"e6c57d77-ac27-46e2-958d-a1598a5de972","Type":"ContainerDied","Data":"802f07b473bd905d0a8e65178e053b02b9493aa4366a5aa15f4dcf7eed36f166"} Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.813084 4858 scope.go:117] "RemoveContainer" containerID="4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.813124 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ns6qw" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.841526 4858 scope.go:117] "RemoveContainer" containerID="0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.860715 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ns6qw"] Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.880926 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ns6qw"] Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.891071 4858 scope.go:117] "RemoveContainer" containerID="302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.920558 4858 scope.go:117] "RemoveContainer" containerID="4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2" Mar 20 09:45:40 crc kubenswrapper[4858]: E0320 09:45:40.921026 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2\": container with ID starting with 4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2 not found: ID does not exist" containerID="4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.921074 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2"} err="failed to get container status \"4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2\": rpc error: code = NotFound desc = could not find container \"4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2\": container with ID starting with 4dd486648f76aad58012b9f3c32dc814e3a24faf53dfc445d8048f38d9c5dcf2 not found: ID does not exist" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.921106 4858 scope.go:117] "RemoveContainer" containerID="0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc" Mar 20 09:45:40 crc kubenswrapper[4858]: E0320 09:45:40.921591 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc\": container with ID starting with 0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc not found: ID does not exist" containerID="0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.921626 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc"} err="failed to get container status \"0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc\": rpc error: code = NotFound desc = could not find container \"0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc\": container with ID starting with 0a415ce017880070a41b32f70944f3e198c5174394247a791192c9f26045b7bc not found: ID does not exist" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.921661 4858 scope.go:117] "RemoveContainer" containerID="302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb" Mar 20 09:45:40 crc kubenswrapper[4858]: E0320 09:45:40.924485 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb\": container with ID starting with 302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb not found: ID does not exist" containerID="302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb" Mar 20 09:45:40 crc kubenswrapper[4858]: I0320 09:45:40.924537 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb"} err="failed to get container status \"302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb\": rpc error: code = NotFound desc = could not find container \"302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb\": container with ID starting with 302bb0cb64d29af40419c8dc1924e551f75a9f0a343f7c8de0dd880b28573acb not found: ID does not exist" Mar 20 09:45:42 crc kubenswrapper[4858]: I0320 09:45:42.094652 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" path="/var/lib/kubelet/pods/e6c57d77-ac27-46e2-958d-a1598a5de972/volumes" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.150666 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566666-69bjt"] Mar 20 09:46:00 crc kubenswrapper[4858]: E0320 09:46:00.152635 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="registry-server" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.152651 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="registry-server" Mar 20 09:46:00 crc kubenswrapper[4858]: E0320 09:46:00.152668 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="extract-content" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.152676 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="extract-content" Mar 20 09:46:00 crc kubenswrapper[4858]: E0320 09:46:00.152693 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="extract-utilities" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.152700 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="extract-utilities" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.152852 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6c57d77-ac27-46e2-958d-a1598a5de972" containerName="registry-server" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.154058 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-69bjt" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.157199 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.157588 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.159419 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.165421 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-69bjt"] Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.266593 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5sb\" (UniqueName: \"kubernetes.io/projected/72eeae25-efc1-4285-8127-173c644400ab-kube-api-access-xm5sb\") pod \"auto-csr-approver-29566666-69bjt\" (UID: \"72eeae25-efc1-4285-8127-173c644400ab\") " pod="openshift-infra/auto-csr-approver-29566666-69bjt" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.367631 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5sb\" (UniqueName: \"kubernetes.io/projected/72eeae25-efc1-4285-8127-173c644400ab-kube-api-access-xm5sb\") pod \"auto-csr-approver-29566666-69bjt\" (UID: \"72eeae25-efc1-4285-8127-173c644400ab\") " pod="openshift-infra/auto-csr-approver-29566666-69bjt" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.396876 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5sb\" (UniqueName: \"kubernetes.io/projected/72eeae25-efc1-4285-8127-173c644400ab-kube-api-access-xm5sb\") pod \"auto-csr-approver-29566666-69bjt\" (UID: \"72eeae25-efc1-4285-8127-173c644400ab\") " pod="openshift-infra/auto-csr-approver-29566666-69bjt" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.510122 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-69bjt" Mar 20 09:46:00 crc kubenswrapper[4858]: I0320 09:46:00.932825 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-69bjt"] Mar 20 09:46:01 crc kubenswrapper[4858]: I0320 09:46:01.000961 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566666-69bjt" event={"ID":"72eeae25-efc1-4285-8127-173c644400ab","Type":"ContainerStarted","Data":"916a627a46ad8028d85a5660e55cf1ec93be542f1ffae90e939f45c7e576ff14"} Mar 20 09:46:03 crc kubenswrapper[4858]: I0320 09:46:03.035580 4858 generic.go:334] "Generic (PLEG): container finished" podID="72eeae25-efc1-4285-8127-173c644400ab" containerID="e96491fec9fad04ed931f8100cf8790cc50bb31682e6425f655591b8a95f30f1" exitCode=0 Mar 20 09:46:03 crc kubenswrapper[4858]: I0320 09:46:03.035641 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566666-69bjt" event={"ID":"72eeae25-efc1-4285-8127-173c644400ab","Type":"ContainerDied","Data":"e96491fec9fad04ed931f8100cf8790cc50bb31682e6425f655591b8a95f30f1"} Mar 20 09:46:04 crc kubenswrapper[4858]: I0320 09:46:04.458990 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-69bjt" Mar 20 09:46:04 crc kubenswrapper[4858]: I0320 09:46:04.632444 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm5sb\" (UniqueName: \"kubernetes.io/projected/72eeae25-efc1-4285-8127-173c644400ab-kube-api-access-xm5sb\") pod \"72eeae25-efc1-4285-8127-173c644400ab\" (UID: \"72eeae25-efc1-4285-8127-173c644400ab\") " Mar 20 09:46:04 crc kubenswrapper[4858]: I0320 09:46:04.641610 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72eeae25-efc1-4285-8127-173c644400ab-kube-api-access-xm5sb" (OuterVolumeSpecName: "kube-api-access-xm5sb") pod "72eeae25-efc1-4285-8127-173c644400ab" (UID: "72eeae25-efc1-4285-8127-173c644400ab"). InnerVolumeSpecName "kube-api-access-xm5sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:46:04 crc kubenswrapper[4858]: I0320 09:46:04.734838 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm5sb\" (UniqueName: \"kubernetes.io/projected/72eeae25-efc1-4285-8127-173c644400ab-kube-api-access-xm5sb\") on node \"crc\" DevicePath \"\"" Mar 20 09:46:05 crc kubenswrapper[4858]: I0320 09:46:05.054011 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566666-69bjt" event={"ID":"72eeae25-efc1-4285-8127-173c644400ab","Type":"ContainerDied","Data":"916a627a46ad8028d85a5660e55cf1ec93be542f1ffae90e939f45c7e576ff14"} Mar 20 09:46:05 crc kubenswrapper[4858]: I0320 09:46:05.054052 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="916a627a46ad8028d85a5660e55cf1ec93be542f1ffae90e939f45c7e576ff14" Mar 20 09:46:05 crc kubenswrapper[4858]: I0320 09:46:05.054115 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566666-69bjt" Mar 20 09:46:05 crc kubenswrapper[4858]: I0320 09:46:05.550569 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-hxxdt"] Mar 20 09:46:05 crc kubenswrapper[4858]: I0320 09:46:05.561507 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566660-hxxdt"] Mar 20 09:46:06 crc kubenswrapper[4858]: I0320 09:46:06.085076 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b75ada44-8bbf-47bd-a20b-eb9d95138980" path="/var/lib/kubelet/pods/b75ada44-8bbf-47bd-a20b-eb9d95138980/volumes" Mar 20 09:46:07 crc kubenswrapper[4858]: I0320 09:46:07.890278 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:46:07 crc kubenswrapper[4858]: I0320 09:46:07.891557 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:46:31 crc kubenswrapper[4858]: I0320 09:46:31.643089 4858 scope.go:117] "RemoveContainer" containerID="714e78437801933d5a346d4ef966b0dff0682fd7acda463de7d70bec21dea9f5" Mar 20 09:46:37 crc kubenswrapper[4858]: I0320 09:46:37.890722 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:46:37 crc kubenswrapper[4858]: I0320 09:46:37.891449 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:47:07 crc kubenswrapper[4858]: I0320 09:47:07.890541 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:47:07 crc kubenswrapper[4858]: I0320 09:47:07.891164 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:47:07 crc kubenswrapper[4858]: I0320 09:47:07.891230 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:47:07 crc kubenswrapper[4858]: I0320 09:47:07.892127 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"021bb0318f7c2654b5dd7a30ce0569091d573b7357c04b6dbe28f1b4c1f439c8"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:47:07 crc kubenswrapper[4858]: I0320 09:47:07.892220 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://021bb0318f7c2654b5dd7a30ce0569091d573b7357c04b6dbe28f1b4c1f439c8" gracePeriod=600 Mar 20 09:47:08 crc kubenswrapper[4858]: I0320 09:47:08.596440 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="021bb0318f7c2654b5dd7a30ce0569091d573b7357c04b6dbe28f1b4c1f439c8" exitCode=0 Mar 20 09:47:08 crc kubenswrapper[4858]: I0320 09:47:08.596509 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"021bb0318f7c2654b5dd7a30ce0569091d573b7357c04b6dbe28f1b4c1f439c8"} Mar 20 09:47:08 crc kubenswrapper[4858]: I0320 09:47:08.596789 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711"} Mar 20 09:47:08 crc kubenswrapper[4858]: I0320 09:47:08.596812 4858 scope.go:117] "RemoveContainer" containerID="dbe3be1d3c9bf07015a78bceda39dcb568516733a6d3d672321cee726becdcf6" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.024279 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zj2g2"] Mar 20 09:47:41 crc kubenswrapper[4858]: E0320 09:47:41.025307 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72eeae25-efc1-4285-8127-173c644400ab" containerName="oc" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.025350 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="72eeae25-efc1-4285-8127-173c644400ab" containerName="oc" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.025603 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="72eeae25-efc1-4285-8127-173c644400ab" containerName="oc" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.027242 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.043574 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zj2g2"] Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.134010 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxpcq\" (UniqueName: \"kubernetes.io/projected/5c3eca77-74cd-4bd8-920a-81c899aa33b9-kube-api-access-lxpcq\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.134140 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-catalog-content\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.134225 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-utilities\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.236023 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxpcq\" (UniqueName: \"kubernetes.io/projected/5c3eca77-74cd-4bd8-920a-81c899aa33b9-kube-api-access-lxpcq\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.236087 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-catalog-content\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.236136 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-utilities\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.236796 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-utilities\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.237131 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-catalog-content\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.262709 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxpcq\" (UniqueName: \"kubernetes.io/projected/5c3eca77-74cd-4bd8-920a-81c899aa33b9-kube-api-access-lxpcq\") pod \"redhat-marketplace-zj2g2\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.368586 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.823034 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zj2g2"] Mar 20 09:47:41 crc kubenswrapper[4858]: I0320 09:47:41.884909 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zj2g2" event={"ID":"5c3eca77-74cd-4bd8-920a-81c899aa33b9","Type":"ContainerStarted","Data":"cb5c8d2abc698a23f21735673540a1a25831c24495a1c2018d622d32997a5b1c"} Mar 20 09:47:42 crc kubenswrapper[4858]: I0320 09:47:42.895009 4858 generic.go:334] "Generic (PLEG): container finished" podID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerID="1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f" exitCode=0 Mar 20 09:47:42 crc kubenswrapper[4858]: I0320 09:47:42.895113 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zj2g2" event={"ID":"5c3eca77-74cd-4bd8-920a-81c899aa33b9","Type":"ContainerDied","Data":"1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f"} Mar 20 09:47:42 crc kubenswrapper[4858]: I0320 09:47:42.898934 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:47:43 crc kubenswrapper[4858]: I0320 09:47:43.906081 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zj2g2" event={"ID":"5c3eca77-74cd-4bd8-920a-81c899aa33b9","Type":"ContainerStarted","Data":"3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb"} Mar 20 09:47:44 crc kubenswrapper[4858]: I0320 09:47:44.916462 4858 generic.go:334] "Generic (PLEG): container finished" podID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerID="3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb" exitCode=0 Mar 20 09:47:44 crc kubenswrapper[4858]: I0320 09:47:44.916606 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zj2g2" event={"ID":"5c3eca77-74cd-4bd8-920a-81c899aa33b9","Type":"ContainerDied","Data":"3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb"} Mar 20 09:47:45 crc kubenswrapper[4858]: I0320 09:47:45.927128 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zj2g2" event={"ID":"5c3eca77-74cd-4bd8-920a-81c899aa33b9","Type":"ContainerStarted","Data":"fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110"} Mar 20 09:47:45 crc kubenswrapper[4858]: I0320 09:47:45.954445 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zj2g2" podStartSLOduration=2.368097523 podStartE2EDuration="4.954426788s" podCreationTimestamp="2026-03-20 09:47:41 +0000 UTC" firstStartedPulling="2026-03-20 09:47:42.898289927 +0000 UTC m=+3044.218708174" lastFinishedPulling="2026-03-20 09:47:45.484619212 +0000 UTC m=+3046.805037439" observedRunningTime="2026-03-20 09:47:45.947939789 +0000 UTC m=+3047.268358016" watchObservedRunningTime="2026-03-20 09:47:45.954426788 +0000 UTC m=+3047.274844995" Mar 20 09:47:51 crc kubenswrapper[4858]: I0320 09:47:51.369651 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:51 crc kubenswrapper[4858]: I0320 09:47:51.370448 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:51 crc kubenswrapper[4858]: I0320 09:47:51.439820 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:52 crc kubenswrapper[4858]: I0320 09:47:52.046222 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:52 crc kubenswrapper[4858]: I0320 09:47:52.105485 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zj2g2"] Mar 20 09:47:53 crc kubenswrapper[4858]: I0320 09:47:53.993911 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zj2g2" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="registry-server" containerID="cri-o://fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110" gracePeriod=2 Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.562916 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.653284 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxpcq\" (UniqueName: \"kubernetes.io/projected/5c3eca77-74cd-4bd8-920a-81c899aa33b9-kube-api-access-lxpcq\") pod \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.653424 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-catalog-content\") pod \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.653606 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-utilities\") pod \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\" (UID: \"5c3eca77-74cd-4bd8-920a-81c899aa33b9\") " Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.654519 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-utilities" (OuterVolumeSpecName: "utilities") pod "5c3eca77-74cd-4bd8-920a-81c899aa33b9" (UID: "5c3eca77-74cd-4bd8-920a-81c899aa33b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.659827 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c3eca77-74cd-4bd8-920a-81c899aa33b9-kube-api-access-lxpcq" (OuterVolumeSpecName: "kube-api-access-lxpcq") pod "5c3eca77-74cd-4bd8-920a-81c899aa33b9" (UID: "5c3eca77-74cd-4bd8-920a-81c899aa33b9"). InnerVolumeSpecName "kube-api-access-lxpcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.679296 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c3eca77-74cd-4bd8-920a-81c899aa33b9" (UID: "5c3eca77-74cd-4bd8-920a-81c899aa33b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.755032 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.755070 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c3eca77-74cd-4bd8-920a-81c899aa33b9-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:47:54 crc kubenswrapper[4858]: I0320 09:47:54.755082 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxpcq\" (UniqueName: \"kubernetes.io/projected/5c3eca77-74cd-4bd8-920a-81c899aa33b9-kube-api-access-lxpcq\") on node \"crc\" DevicePath \"\"" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.010886 4858 generic.go:334] "Generic (PLEG): container finished" podID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerID="fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110" exitCode=0 Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.010942 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zj2g2" event={"ID":"5c3eca77-74cd-4bd8-920a-81c899aa33b9","Type":"ContainerDied","Data":"fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110"} Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.010972 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zj2g2" event={"ID":"5c3eca77-74cd-4bd8-920a-81c899aa33b9","Type":"ContainerDied","Data":"cb5c8d2abc698a23f21735673540a1a25831c24495a1c2018d622d32997a5b1c"} Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.010992 4858 scope.go:117] "RemoveContainer" containerID="fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.011159 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zj2g2" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.056284 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zj2g2"] Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.063965 4858 scope.go:117] "RemoveContainer" containerID="3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.066834 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zj2g2"] Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.095777 4858 scope.go:117] "RemoveContainer" containerID="1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.140117 4858 scope.go:117] "RemoveContainer" containerID="fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110" Mar 20 09:47:55 crc kubenswrapper[4858]: E0320 09:47:55.140577 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110\": container with ID starting with fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110 not found: ID does not exist" containerID="fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.140626 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110"} err="failed to get container status \"fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110\": rpc error: code = NotFound desc = could not find container \"fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110\": container with ID starting with fc5b743a9cca932f0da9301f3b683986b6325e402f39c764d3b26d920411c110 not found: ID does not exist" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.140659 4858 scope.go:117] "RemoveContainer" containerID="3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb" Mar 20 09:47:55 crc kubenswrapper[4858]: E0320 09:47:55.141131 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb\": container with ID starting with 3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb not found: ID does not exist" containerID="3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.141172 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb"} err="failed to get container status \"3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb\": rpc error: code = NotFound desc = could not find container \"3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb\": container with ID starting with 3d45ae1185a000d90b31ec34febe14abefdf4aaee4cedef5d6b62a738916bbdb not found: ID does not exist" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.141201 4858 scope.go:117] "RemoveContainer" containerID="1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f" Mar 20 09:47:55 crc kubenswrapper[4858]: E0320 09:47:55.141901 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f\": container with ID starting with 1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f not found: ID does not exist" containerID="1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f" Mar 20 09:47:55 crc kubenswrapper[4858]: I0320 09:47:55.141946 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f"} err="failed to get container status \"1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f\": rpc error: code = NotFound desc = could not find container \"1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f\": container with ID starting with 1f196d2b321a7a19da5bd3a02e57a1510ab3846db2da1b25ef18df6bf4cac48f not found: ID does not exist" Mar 20 09:47:56 crc kubenswrapper[4858]: I0320 09:47:56.087052 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" path="/var/lib/kubelet/pods/5c3eca77-74cd-4bd8-920a-81c899aa33b9/volumes" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.150929 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566668-n2gcc"] Mar 20 09:48:00 crc kubenswrapper[4858]: E0320 09:48:00.152067 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="extract-utilities" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.152083 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="extract-utilities" Mar 20 09:48:00 crc kubenswrapper[4858]: E0320 09:48:00.152105 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="registry-server" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.152112 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="registry-server" Mar 20 09:48:00 crc kubenswrapper[4858]: E0320 09:48:00.152122 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="extract-content" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.152129 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="extract-content" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.152267 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c3eca77-74cd-4bd8-920a-81c899aa33b9" containerName="registry-server" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.152946 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-n2gcc" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.156912 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.157123 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.157288 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.168067 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566668-n2gcc"] Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.238732 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrkn7\" (UniqueName: \"kubernetes.io/projected/1cdceba8-aa1f-4931-930f-612d6ca0161d-kube-api-access-lrkn7\") pod \"auto-csr-approver-29566668-n2gcc\" (UID: \"1cdceba8-aa1f-4931-930f-612d6ca0161d\") " pod="openshift-infra/auto-csr-approver-29566668-n2gcc" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.340420 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrkn7\" (UniqueName: \"kubernetes.io/projected/1cdceba8-aa1f-4931-930f-612d6ca0161d-kube-api-access-lrkn7\") pod \"auto-csr-approver-29566668-n2gcc\" (UID: \"1cdceba8-aa1f-4931-930f-612d6ca0161d\") " pod="openshift-infra/auto-csr-approver-29566668-n2gcc" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.360906 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrkn7\" (UniqueName: \"kubernetes.io/projected/1cdceba8-aa1f-4931-930f-612d6ca0161d-kube-api-access-lrkn7\") pod \"auto-csr-approver-29566668-n2gcc\" (UID: \"1cdceba8-aa1f-4931-930f-612d6ca0161d\") " pod="openshift-infra/auto-csr-approver-29566668-n2gcc" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.484770 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-n2gcc" Mar 20 09:48:00 crc kubenswrapper[4858]: I0320 09:48:00.929627 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566668-n2gcc"] Mar 20 09:48:01 crc kubenswrapper[4858]: I0320 09:48:01.063062 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566668-n2gcc" event={"ID":"1cdceba8-aa1f-4931-930f-612d6ca0161d","Type":"ContainerStarted","Data":"173271fcdcc7f4751f827aa0dcb3d8d1badf3329e77d4ab5a7af5bc3f2289f5a"} Mar 20 09:48:03 crc kubenswrapper[4858]: I0320 09:48:03.087558 4858 generic.go:334] "Generic (PLEG): container finished" podID="1cdceba8-aa1f-4931-930f-612d6ca0161d" containerID="cff5d1b3d6c4f1e18e142ef5b564e2ea4cab3e585dc82acaa59ff109c856cb2d" exitCode=0 Mar 20 09:48:03 crc kubenswrapper[4858]: I0320 09:48:03.087695 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566668-n2gcc" event={"ID":"1cdceba8-aa1f-4931-930f-612d6ca0161d","Type":"ContainerDied","Data":"cff5d1b3d6c4f1e18e142ef5b564e2ea4cab3e585dc82acaa59ff109c856cb2d"} Mar 20 09:48:04 crc kubenswrapper[4858]: I0320 09:48:04.465074 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-n2gcc" Mar 20 09:48:04 crc kubenswrapper[4858]: I0320 09:48:04.503972 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrkn7\" (UniqueName: \"kubernetes.io/projected/1cdceba8-aa1f-4931-930f-612d6ca0161d-kube-api-access-lrkn7\") pod \"1cdceba8-aa1f-4931-930f-612d6ca0161d\" (UID: \"1cdceba8-aa1f-4931-930f-612d6ca0161d\") " Mar 20 09:48:04 crc kubenswrapper[4858]: I0320 09:48:04.510523 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cdceba8-aa1f-4931-930f-612d6ca0161d-kube-api-access-lrkn7" (OuterVolumeSpecName: "kube-api-access-lrkn7") pod "1cdceba8-aa1f-4931-930f-612d6ca0161d" (UID: "1cdceba8-aa1f-4931-930f-612d6ca0161d"). InnerVolumeSpecName "kube-api-access-lrkn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:48:04 crc kubenswrapper[4858]: I0320 09:48:04.606069 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrkn7\" (UniqueName: \"kubernetes.io/projected/1cdceba8-aa1f-4931-930f-612d6ca0161d-kube-api-access-lrkn7\") on node \"crc\" DevicePath \"\"" Mar 20 09:48:05 crc kubenswrapper[4858]: I0320 09:48:05.104335 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566668-n2gcc" event={"ID":"1cdceba8-aa1f-4931-930f-612d6ca0161d","Type":"ContainerDied","Data":"173271fcdcc7f4751f827aa0dcb3d8d1badf3329e77d4ab5a7af5bc3f2289f5a"} Mar 20 09:48:05 crc kubenswrapper[4858]: I0320 09:48:05.104378 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="173271fcdcc7f4751f827aa0dcb3d8d1badf3329e77d4ab5a7af5bc3f2289f5a" Mar 20 09:48:05 crc kubenswrapper[4858]: I0320 09:48:05.104420 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566668-n2gcc" Mar 20 09:48:05 crc kubenswrapper[4858]: I0320 09:48:05.541004 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-gx9st"] Mar 20 09:48:05 crc kubenswrapper[4858]: I0320 09:48:05.549148 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566662-gx9st"] Mar 20 09:48:06 crc kubenswrapper[4858]: I0320 09:48:06.086665 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc444685-16df-4295-b62a-c0d9e26af1d7" path="/var/lib/kubelet/pods/cc444685-16df-4295-b62a-c0d9e26af1d7/volumes" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.106821 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-m7hst/must-gather-lqx8h"] Mar 20 09:48:24 crc kubenswrapper[4858]: E0320 09:48:24.107930 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdceba8-aa1f-4931-930f-612d6ca0161d" containerName="oc" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.107943 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdceba8-aa1f-4931-930f-612d6ca0161d" containerName="oc" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.108093 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cdceba8-aa1f-4931-930f-612d6ca0161d" containerName="oc" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.108917 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.119000 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m7hst"/"kube-root-ca.crt" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.119568 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-m7hst"/"openshift-service-ca.crt" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.127804 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-m7hst"/"default-dockercfg-knlzn" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.139229 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m7hst/must-gather-lqx8h"] Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.198182 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60bf6a80-0c17-41d8-a4af-46c4dc800572-must-gather-output\") pod \"must-gather-lqx8h\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.198289 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjbhw\" (UniqueName: \"kubernetes.io/projected/60bf6a80-0c17-41d8-a4af-46c4dc800572-kube-api-access-mjbhw\") pod \"must-gather-lqx8h\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.300016 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60bf6a80-0c17-41d8-a4af-46c4dc800572-must-gather-output\") pod \"must-gather-lqx8h\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.300071 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjbhw\" (UniqueName: \"kubernetes.io/projected/60bf6a80-0c17-41d8-a4af-46c4dc800572-kube-api-access-mjbhw\") pod \"must-gather-lqx8h\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.300722 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60bf6a80-0c17-41d8-a4af-46c4dc800572-must-gather-output\") pod \"must-gather-lqx8h\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.317751 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjbhw\" (UniqueName: \"kubernetes.io/projected/60bf6a80-0c17-41d8-a4af-46c4dc800572-kube-api-access-mjbhw\") pod \"must-gather-lqx8h\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.425893 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:48:24 crc kubenswrapper[4858]: I0320 09:48:24.864624 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-m7hst/must-gather-lqx8h"] Mar 20 09:48:25 crc kubenswrapper[4858]: I0320 09:48:25.260775 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m7hst/must-gather-lqx8h" event={"ID":"60bf6a80-0c17-41d8-a4af-46c4dc800572","Type":"ContainerStarted","Data":"aeb98f709b336decfc59ace598d317200e552dc8de3d702a90a6d9df081775c1"} Mar 20 09:48:31 crc kubenswrapper[4858]: I0320 09:48:31.359237 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m7hst/must-gather-lqx8h" event={"ID":"60bf6a80-0c17-41d8-a4af-46c4dc800572","Type":"ContainerStarted","Data":"cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a"} Mar 20 09:48:31 crc kubenswrapper[4858]: I0320 09:48:31.751637 4858 scope.go:117] "RemoveContainer" containerID="ebca739e8fd0b796211a78f1349bf65fe26e90053909bfe64d7549f0fd74e392" Mar 20 09:48:32 crc kubenswrapper[4858]: I0320 09:48:32.368217 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m7hst/must-gather-lqx8h" event={"ID":"60bf6a80-0c17-41d8-a4af-46c4dc800572","Type":"ContainerStarted","Data":"3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1"} Mar 20 09:49:20 crc kubenswrapper[4858]: I0320 09:49:20.893482 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-tgjql_55debadc-f9be-4dc0-a269-4e8782024065/init/0.log" Mar 20 09:49:21 crc kubenswrapper[4858]: I0320 09:49:21.132238 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-tgjql_55debadc-f9be-4dc0-a269-4e8782024065/init/0.log" Mar 20 09:49:21 crc kubenswrapper[4858]: I0320 09:49:21.152234 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78dd6ddcc-tgjql_55debadc-f9be-4dc0-a269-4e8782024065/dnsmasq-dns/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.048104 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph_6c4644e0-05b7-4776-b0ae-d45502e6f6b4/util/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.224093 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph_6c4644e0-05b7-4776-b0ae-d45502e6f6b4/pull/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.252336 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph_6c4644e0-05b7-4776-b0ae-d45502e6f6b4/util/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.272045 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph_6c4644e0-05b7-4776-b0ae-d45502e6f6b4/pull/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.423645 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph_6c4644e0-05b7-4776-b0ae-d45502e6f6b4/pull/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.441784 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph_6c4644e0-05b7-4776-b0ae-d45502e6f6b4/extract/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.442790 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b4a728943c10eb3a8d997059249394951a4c19e9bf2c22f3d6025f4badddfph_6c4644e0-05b7-4776-b0ae-d45502e6f6b4/util/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.616887 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-zk2d4_0e2d5479-7826-4759-99ed-c3775a01035c/manager/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.867538 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-8bwmb_e2a119a6-367e-4a85-8617-e83d9f6a23b7/manager/0.log" Mar 20 09:49:36 crc kubenswrapper[4858]: I0320 09:49:36.970905 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-bb9mf_81033dc9-91dc-4bff-b69d-e1171f82b83c/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.085470 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-ng56m_feae143f-10cc-4412-9c35-70499e77b5bd/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.089402 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-7c7lv_cffb3d94-5d49-4a09-a1d8-d41e15cb9e6e/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.280208 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-sk6sh_1e41ad93-0360-4a67-a155-ee05abfbcee1/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.293071 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-577ccd856-nw2q9_a0e45ec4-0059-41b8-8897-8004d4adb9da/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.396860 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-cb6dz_1adeed8a-d531-46ee-b037-ea137468e026/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.449331 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-bl6x4_3dfa5be0-6028-4091-8129-71fa07ab93ac/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.555003 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-sv6f4_71693924-c907-41df-b4ab-cbd9bfb7f97d/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.615580 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-4pprb_1435162c-2ff0-4013-b838-48b63a57933b/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.752096 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-n4k5l_44f13e51-a8b3-489d-a4bb-852a441759f8/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.824094 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-sxlpj_213a0cad-a19e-4337-ab03-67e0cc63fa08/manager/0.log" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.890222 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.890571 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:49:37 crc kubenswrapper[4858]: I0320 09:49:37.929646 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-tw6tg_8aa33b96-bd71-46c2-814a-ade6d5181f8a/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.046642 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-9gjsl_9e387ade-406a-4372-a097-554d1572296c/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.195002 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-9df8dd5fd-7ssln_23fed3a3-2d6d-4c5a-9354-ac8f4b25f2de/operator/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.334137 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-55958644c4-zzhvj_4c0a5337-e5f9-4ac0-8123-e7cfd4fceea7/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.370452 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-s44rj_688b14a3-bd9d-45d9-8adc-c65703cfcd9d/registry-server/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.508866 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-cwdpq_87d69191-f1ca-46ea-a082-fdc249e342e1/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.588713 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-kvg89_eeb95c09-093f-48d0-8a80-d52fc8cb7157/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.660473 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-rbl7w_ab86f223-d7e2-4527-9b1b-eb8633bafd01/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.761526 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-86cpb_a5c893bf-acfe-4ec9-a63a-3c48055530f4/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.801179 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-zgtwl_d4c562cc-1a1b-4c09-b468-314dd82774f3/manager/0.log" Mar 20 09:49:38 crc kubenswrapper[4858]: I0320 09:49:38.924993 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-zgbc2_b91d223a-93ba-4db6-8153-9388e0e8a3a4/manager/0.log" Mar 20 09:49:58 crc kubenswrapper[4858]: I0320 09:49:58.373766 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-msrvs_b51deb4e-ca50-41d4-8b00-bb996f8e7782/control-plane-machine-set-operator/0.log" Mar 20 09:49:58 crc kubenswrapper[4858]: I0320 09:49:58.572291 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vw55r_3c53fc26-4e6d-4d8f-bb46-59987bcc746f/kube-rbac-proxy/0.log" Mar 20 09:49:58 crc kubenswrapper[4858]: I0320 09:49:58.572735 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vw55r_3c53fc26-4e6d-4d8f-bb46-59987bcc746f/machine-api-operator/0.log" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.136377 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-m7hst/must-gather-lqx8h" podStartSLOduration=90.075317419 podStartE2EDuration="1m36.136352739s" podCreationTimestamp="2026-03-20 09:48:24 +0000 UTC" firstStartedPulling="2026-03-20 09:48:24.873810377 +0000 UTC m=+3086.194228584" lastFinishedPulling="2026-03-20 09:48:30.934845697 +0000 UTC m=+3092.255263904" observedRunningTime="2026-03-20 09:48:32.404249017 +0000 UTC m=+3093.724667214" watchObservedRunningTime="2026-03-20 09:50:00.136352739 +0000 UTC m=+3181.456770966" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.143127 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566670-6qb6v"] Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.144831 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-6qb6v" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.146744 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.146894 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.147233 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.152575 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566670-6qb6v"] Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.203496 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw8hh\" (UniqueName: \"kubernetes.io/projected/e6bb9eff-c7c8-4754-a74a-80e52fd3f103-kube-api-access-nw8hh\") pod \"auto-csr-approver-29566670-6qb6v\" (UID: \"e6bb9eff-c7c8-4754-a74a-80e52fd3f103\") " pod="openshift-infra/auto-csr-approver-29566670-6qb6v" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.304812 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw8hh\" (UniqueName: \"kubernetes.io/projected/e6bb9eff-c7c8-4754-a74a-80e52fd3f103-kube-api-access-nw8hh\") pod \"auto-csr-approver-29566670-6qb6v\" (UID: \"e6bb9eff-c7c8-4754-a74a-80e52fd3f103\") " pod="openshift-infra/auto-csr-approver-29566670-6qb6v" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.343894 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw8hh\" (UniqueName: \"kubernetes.io/projected/e6bb9eff-c7c8-4754-a74a-80e52fd3f103-kube-api-access-nw8hh\") pod \"auto-csr-approver-29566670-6qb6v\" (UID: \"e6bb9eff-c7c8-4754-a74a-80e52fd3f103\") " pod="openshift-infra/auto-csr-approver-29566670-6qb6v" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.472872 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-6qb6v" Mar 20 09:50:00 crc kubenswrapper[4858]: I0320 09:50:00.940167 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566670-6qb6v"] Mar 20 09:50:01 crc kubenswrapper[4858]: I0320 09:50:01.057224 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566670-6qb6v" event={"ID":"e6bb9eff-c7c8-4754-a74a-80e52fd3f103","Type":"ContainerStarted","Data":"07fc4710d7bba1142bfbea7231dad4c9aee7dc1228beeef6ba5ec6dae8e4519d"} Mar 20 09:50:03 crc kubenswrapper[4858]: I0320 09:50:03.088925 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566670-6qb6v" event={"ID":"e6bb9eff-c7c8-4754-a74a-80e52fd3f103","Type":"ContainerDied","Data":"c40f242b9627890fe01afcb61632912d0c98c2014bec1c53895949ff2dbca70a"} Mar 20 09:50:03 crc kubenswrapper[4858]: I0320 09:50:03.088942 4858 generic.go:334] "Generic (PLEG): container finished" podID="e6bb9eff-c7c8-4754-a74a-80e52fd3f103" containerID="c40f242b9627890fe01afcb61632912d0c98c2014bec1c53895949ff2dbca70a" exitCode=0 Mar 20 09:50:04 crc kubenswrapper[4858]: I0320 09:50:04.373211 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-6qb6v" Mar 20 09:50:04 crc kubenswrapper[4858]: I0320 09:50:04.569019 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw8hh\" (UniqueName: \"kubernetes.io/projected/e6bb9eff-c7c8-4754-a74a-80e52fd3f103-kube-api-access-nw8hh\") pod \"e6bb9eff-c7c8-4754-a74a-80e52fd3f103\" (UID: \"e6bb9eff-c7c8-4754-a74a-80e52fd3f103\") " Mar 20 09:50:04 crc kubenswrapper[4858]: I0320 09:50:04.578568 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6bb9eff-c7c8-4754-a74a-80e52fd3f103-kube-api-access-nw8hh" (OuterVolumeSpecName: "kube-api-access-nw8hh") pod "e6bb9eff-c7c8-4754-a74a-80e52fd3f103" (UID: "e6bb9eff-c7c8-4754-a74a-80e52fd3f103"). InnerVolumeSpecName "kube-api-access-nw8hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:50:04 crc kubenswrapper[4858]: I0320 09:50:04.671767 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw8hh\" (UniqueName: \"kubernetes.io/projected/e6bb9eff-c7c8-4754-a74a-80e52fd3f103-kube-api-access-nw8hh\") on node \"crc\" DevicePath \"\"" Mar 20 09:50:05 crc kubenswrapper[4858]: I0320 09:50:05.106116 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566670-6qb6v" event={"ID":"e6bb9eff-c7c8-4754-a74a-80e52fd3f103","Type":"ContainerDied","Data":"07fc4710d7bba1142bfbea7231dad4c9aee7dc1228beeef6ba5ec6dae8e4519d"} Mar 20 09:50:05 crc kubenswrapper[4858]: I0320 09:50:05.106159 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07fc4710d7bba1142bfbea7231dad4c9aee7dc1228beeef6ba5ec6dae8e4519d" Mar 20 09:50:05 crc kubenswrapper[4858]: I0320 09:50:05.106188 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566670-6qb6v" Mar 20 09:50:05 crc kubenswrapper[4858]: I0320 09:50:05.452679 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-g4tn4"] Mar 20 09:50:05 crc kubenswrapper[4858]: I0320 09:50:05.457390 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566664-g4tn4"] Mar 20 09:50:06 crc kubenswrapper[4858]: I0320 09:50:06.083666 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="065a313a-b2c9-4b27-b063-bd71ed126811" path="/var/lib/kubelet/pods/065a313a-b2c9-4b27-b063-bd71ed126811/volumes" Mar 20 09:50:07 crc kubenswrapper[4858]: I0320 09:50:07.890238 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:50:07 crc kubenswrapper[4858]: I0320 09:50:07.890309 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:50:11 crc kubenswrapper[4858]: I0320 09:50:11.654932 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-s5c5l_714bde66-3be8-485f-af48-b32f2177ae05/cert-manager-controller/0.log" Mar 20 09:50:11 crc kubenswrapper[4858]: I0320 09:50:11.778906 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-zxsx6_e7decc95-2f64-45ce-a3d5-679589f5b77a/cert-manager-cainjector/0.log" Mar 20 09:50:11 crc kubenswrapper[4858]: I0320 09:50:11.849173 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-dgqzw_76d3df4f-4c5f-45ad-8cb4-4f11a7821fb3/cert-manager-webhook/0.log" Mar 20 09:50:25 crc kubenswrapper[4858]: I0320 09:50:25.446566 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-njbj4_7a21679b-2a77-4328-9648-14933286fb41/nmstate-console-plugin/0.log" Mar 20 09:50:25 crc kubenswrapper[4858]: I0320 09:50:25.580719 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-b2pcs_5101bce4-d1bf-478e-82da-449ec0f98fca/nmstate-handler/0.log" Mar 20 09:50:25 crc kubenswrapper[4858]: I0320 09:50:25.635876 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-6jqct_f652e520-5e0b-4479-9b9b-c4abdc2c27a9/kube-rbac-proxy/0.log" Mar 20 09:50:25 crc kubenswrapper[4858]: I0320 09:50:25.668017 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-6jqct_f652e520-5e0b-4479-9b9b-c4abdc2c27a9/nmstate-metrics/0.log" Mar 20 09:50:25 crc kubenswrapper[4858]: I0320 09:50:25.855778 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-hkqkj_49466b7e-1091-4d2f-9b3c-863941f4744d/nmstate-webhook/0.log" Mar 20 09:50:25 crc kubenswrapper[4858]: I0320 09:50:25.866111 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-q94l9_2105dfd4-78ef-4fd6-a179-02ad553bef8f/nmstate-operator/0.log" Mar 20 09:50:31 crc kubenswrapper[4858]: I0320 09:50:31.847103 4858 scope.go:117] "RemoveContainer" containerID="47cba87d5c4976d6c4bf65ca3efe1bc5c981ce059cea321eb3a1c12d6ba8d8c3" Mar 20 09:50:37 crc kubenswrapper[4858]: I0320 09:50:37.890682 4858 patch_prober.go:28] interesting pod/machine-config-daemon-w6t79 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 20 09:50:37 crc kubenswrapper[4858]: I0320 09:50:37.891018 4858 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 20 09:50:37 crc kubenswrapper[4858]: I0320 09:50:37.891063 4858 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" Mar 20 09:50:37 crc kubenswrapper[4858]: I0320 09:50:37.891722 4858 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711"} pod="openshift-machine-config-operator/machine-config-daemon-w6t79" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 20 09:50:37 crc kubenswrapper[4858]: I0320 09:50:37.891770 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" containerName="machine-config-daemon" containerID="cri-o://393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" gracePeriod=600 Mar 20 09:50:38 crc kubenswrapper[4858]: E0320 09:50:38.018004 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:50:38 crc kubenswrapper[4858]: I0320 09:50:38.869529 4858 generic.go:334] "Generic (PLEG): container finished" podID="584bd2e0-0786-4137-9674-790c8fb680c5" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" exitCode=0 Mar 20 09:50:38 crc kubenswrapper[4858]: I0320 09:50:38.869643 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerDied","Data":"393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711"} Mar 20 09:50:38 crc kubenswrapper[4858]: I0320 09:50:38.869871 4858 scope.go:117] "RemoveContainer" containerID="021bb0318f7c2654b5dd7a30ce0569091d573b7357c04b6dbe28f1b4c1f439c8" Mar 20 09:50:38 crc kubenswrapper[4858]: I0320 09:50:38.870528 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:50:38 crc kubenswrapper[4858]: E0320 09:50:38.870915 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:50:53 crc kubenswrapper[4858]: I0320 09:50:53.070798 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:50:53 crc kubenswrapper[4858]: E0320 09:50:53.071443 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:50:56 crc kubenswrapper[4858]: I0320 09:50:56.763717 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-dm6h9_e548e45c-1cb2-48a4-bc75-679f254219e5/kube-rbac-proxy/0.log" Mar 20 09:50:56 crc kubenswrapper[4858]: I0320 09:50:56.811619 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-dm6h9_e548e45c-1cb2-48a4-bc75-679f254219e5/controller/0.log" Mar 20 09:50:56 crc kubenswrapper[4858]: I0320 09:50:56.954007 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-frr-files/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.173835 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-reloader/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.175680 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-reloader/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.179040 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-metrics/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.192557 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-frr-files/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.341130 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-metrics/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.357055 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-frr-files/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.365491 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-metrics/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.365655 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-reloader/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.540798 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-reloader/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.568445 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-metrics/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.586843 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/cp-frr-files/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.607181 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/controller/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.762044 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/frr-metrics/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.764577 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/kube-rbac-proxy/0.log" Mar 20 09:50:57 crc kubenswrapper[4858]: I0320 09:50:57.806531 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/kube-rbac-proxy-frr/0.log" Mar 20 09:50:58 crc kubenswrapper[4858]: I0320 09:50:58.011497 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/reloader/0.log" Mar 20 09:50:58 crc kubenswrapper[4858]: I0320 09:50:58.044959 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-2jb95_df73c110-c68f-4dd3-b70b-e0898869c0a6/frr-k8s-webhook-server/0.log" Mar 20 09:50:58 crc kubenswrapper[4858]: I0320 09:50:58.202486 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-79f996c48b-5x54t_f8fa6d2d-8e34-4685-8517-92ffa49a5dcd/manager/0.log" Mar 20 09:50:58 crc kubenswrapper[4858]: I0320 09:50:58.229500 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lwq5q_62cd2367-51df-4a29-a892-cb01bbf2a98b/frr/0.log" Mar 20 09:50:58 crc kubenswrapper[4858]: I0320 09:50:58.382945 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h4sxs_b2351ddb-14a8-445f-9326-9e49d955e417/kube-rbac-proxy/0.log" Mar 20 09:50:58 crc kubenswrapper[4858]: I0320 09:50:58.404576 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-574b899bbf-vpkw6_cc614ec0-61c6-42db-9c82-17f245a66dbb/webhook-server/0.log" Mar 20 09:50:58 crc kubenswrapper[4858]: I0320 09:50:58.517779 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-h4sxs_b2351ddb-14a8-445f-9326-9e49d955e417/speaker/0.log" Mar 20 09:51:07 crc kubenswrapper[4858]: I0320 09:51:07.070568 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:51:07 crc kubenswrapper[4858]: E0320 09:51:07.071533 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:51:11 crc kubenswrapper[4858]: I0320 09:51:11.655347 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg_20a682c7-79f0-4bf3-8574-4beee0f0415f/util/0.log" Mar 20 09:51:11 crc kubenswrapper[4858]: I0320 09:51:11.791083 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg_20a682c7-79f0-4bf3-8574-4beee0f0415f/util/0.log" Mar 20 09:51:11 crc kubenswrapper[4858]: I0320 09:51:11.823198 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg_20a682c7-79f0-4bf3-8574-4beee0f0415f/pull/0.log" Mar 20 09:51:11 crc kubenswrapper[4858]: I0320 09:51:11.827069 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg_20a682c7-79f0-4bf3-8574-4beee0f0415f/pull/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.005411 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg_20a682c7-79f0-4bf3-8574-4beee0f0415f/util/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.014199 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg_20a682c7-79f0-4bf3-8574-4beee0f0415f/extract/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.034684 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874s5gbg_20a682c7-79f0-4bf3-8574-4beee0f0415f/pull/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.191230 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6_59f5aa23-5099-460c-8904-d9bf6acd9958/util/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.355901 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6_59f5aa23-5099-460c-8904-d9bf6acd9958/util/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.362734 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6_59f5aa23-5099-460c-8904-d9bf6acd9958/pull/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.430398 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6_59f5aa23-5099-460c-8904-d9bf6acd9958/pull/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.526668 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6_59f5aa23-5099-460c-8904-d9bf6acd9958/pull/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.545080 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6_59f5aa23-5099-460c-8904-d9bf6acd9958/util/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.571167 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1g6jq6_59f5aa23-5099-460c-8904-d9bf6acd9958/extract/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.697055 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d_8e5953bd-8302-4918-a2da-911d68e7fc79/util/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.874040 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d_8e5953bd-8302-4918-a2da-911d68e7fc79/pull/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.874066 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d_8e5953bd-8302-4918-a2da-911d68e7fc79/util/0.log" Mar 20 09:51:12 crc kubenswrapper[4858]: I0320 09:51:12.939375 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d_8e5953bd-8302-4918-a2da-911d68e7fc79/pull/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.052963 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d_8e5953bd-8302-4918-a2da-911d68e7fc79/util/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.055700 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d_8e5953bd-8302-4918-a2da-911d68e7fc79/extract/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.106756 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dfn2d_8e5953bd-8302-4918-a2da-911d68e7fc79/pull/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.239391 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d65zd_1b309824-10a9-4914-bcc2-e6ec55e6da20/extract-utilities/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.372217 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d65zd_1b309824-10a9-4914-bcc2-e6ec55e6da20/extract-utilities/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.409073 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d65zd_1b309824-10a9-4914-bcc2-e6ec55e6da20/extract-content/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.454867 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d65zd_1b309824-10a9-4914-bcc2-e6ec55e6da20/extract-content/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.560991 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d65zd_1b309824-10a9-4914-bcc2-e6ec55e6da20/extract-utilities/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.561568 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d65zd_1b309824-10a9-4914-bcc2-e6ec55e6da20/extract-content/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.743070 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tnvtl_194ca060-bd99-421b-ac6d-884d999bf54d/extract-utilities/0.log" Mar 20 09:51:13 crc kubenswrapper[4858]: I0320 09:51:13.944777 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tnvtl_194ca060-bd99-421b-ac6d-884d999bf54d/extract-utilities/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.054178 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d65zd_1b309824-10a9-4914-bcc2-e6ec55e6da20/registry-server/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.080391 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tnvtl_194ca060-bd99-421b-ac6d-884d999bf54d/extract-content/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.086583 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tnvtl_194ca060-bd99-421b-ac6d-884d999bf54d/extract-content/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.182604 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tnvtl_194ca060-bd99-421b-ac6d-884d999bf54d/extract-utilities/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.210072 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tnvtl_194ca060-bd99-421b-ac6d-884d999bf54d/extract-content/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.350709 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-st6tw_bbf56bc9-5bfa-4aab-8633-a596385f59a5/marketplace-operator/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.486701 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hkxwl_40519ad0-414d-4c1c-86f1-45ca54a1ab73/extract-utilities/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.652479 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tnvtl_194ca060-bd99-421b-ac6d-884d999bf54d/registry-server/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.667407 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hkxwl_40519ad0-414d-4c1c-86f1-45ca54a1ab73/extract-content/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.692569 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hkxwl_40519ad0-414d-4c1c-86f1-45ca54a1ab73/extract-utilities/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.703104 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hkxwl_40519ad0-414d-4c1c-86f1-45ca54a1ab73/extract-content/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.846104 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hkxwl_40519ad0-414d-4c1c-86f1-45ca54a1ab73/extract-content/0.log" Mar 20 09:51:14 crc kubenswrapper[4858]: I0320 09:51:14.872839 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hkxwl_40519ad0-414d-4c1c-86f1-45ca54a1ab73/extract-utilities/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.011381 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-hkxwl_40519ad0-414d-4c1c-86f1-45ca54a1ab73/registry-server/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.079421 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p9dvb_85554450-9564-4cfa-9410-16dad7d9a3d2/extract-utilities/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.175953 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p9dvb_85554450-9564-4cfa-9410-16dad7d9a3d2/extract-content/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.183945 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p9dvb_85554450-9564-4cfa-9410-16dad7d9a3d2/extract-content/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.184749 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p9dvb_85554450-9564-4cfa-9410-16dad7d9a3d2/extract-utilities/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.363957 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p9dvb_85554450-9564-4cfa-9410-16dad7d9a3d2/extract-content/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.372518 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p9dvb_85554450-9564-4cfa-9410-16dad7d9a3d2/extract-utilities/0.log" Mar 20 09:51:15 crc kubenswrapper[4858]: I0320 09:51:15.703429 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-p9dvb_85554450-9564-4cfa-9410-16dad7d9a3d2/registry-server/0.log" Mar 20 09:51:21 crc kubenswrapper[4858]: I0320 09:51:21.070508 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:51:21 crc kubenswrapper[4858]: E0320 09:51:21.071209 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:51:34 crc kubenswrapper[4858]: I0320 09:51:34.070036 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:51:34 crc kubenswrapper[4858]: E0320 09:51:34.070874 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:51:45 crc kubenswrapper[4858]: I0320 09:51:45.070592 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:51:45 crc kubenswrapper[4858]: E0320 09:51:45.071864 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.078521 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:52:00 crc kubenswrapper[4858]: E0320 09:52:00.079425 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.199905 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566672-gf59n"] Mar 20 09:52:00 crc kubenswrapper[4858]: E0320 09:52:00.200564 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bb9eff-c7c8-4754-a74a-80e52fd3f103" containerName="oc" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.200601 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bb9eff-c7c8-4754-a74a-80e52fd3f103" containerName="oc" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.200872 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bb9eff-c7c8-4754-a74a-80e52fd3f103" containerName="oc" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.201718 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-gf59n" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.204201 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.204519 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.204755 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.207284 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566672-gf59n"] Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.352862 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgddw\" (UniqueName: \"kubernetes.io/projected/5d3050c9-f20b-4277-9629-15c3b95d12c4-kube-api-access-kgddw\") pod \"auto-csr-approver-29566672-gf59n\" (UID: \"5d3050c9-f20b-4277-9629-15c3b95d12c4\") " pod="openshift-infra/auto-csr-approver-29566672-gf59n" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.454433 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgddw\" (UniqueName: \"kubernetes.io/projected/5d3050c9-f20b-4277-9629-15c3b95d12c4-kube-api-access-kgddw\") pod \"auto-csr-approver-29566672-gf59n\" (UID: \"5d3050c9-f20b-4277-9629-15c3b95d12c4\") " pod="openshift-infra/auto-csr-approver-29566672-gf59n" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.492942 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgddw\" (UniqueName: \"kubernetes.io/projected/5d3050c9-f20b-4277-9629-15c3b95d12c4-kube-api-access-kgddw\") pod \"auto-csr-approver-29566672-gf59n\" (UID: \"5d3050c9-f20b-4277-9629-15c3b95d12c4\") " pod="openshift-infra/auto-csr-approver-29566672-gf59n" Mar 20 09:52:00 crc kubenswrapper[4858]: I0320 09:52:00.530621 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-gf59n" Mar 20 09:52:01 crc kubenswrapper[4858]: I0320 09:52:01.028922 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566672-gf59n"] Mar 20 09:52:01 crc kubenswrapper[4858]: I0320 09:52:01.554857 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-gf59n" event={"ID":"5d3050c9-f20b-4277-9629-15c3b95d12c4","Type":"ContainerStarted","Data":"63c570fcf14c481e3ab1693496d298a87eecf1d6083b0fc3573a5c6f4a600b5b"} Mar 20 09:52:02 crc kubenswrapper[4858]: I0320 09:52:02.563039 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-gf59n" event={"ID":"5d3050c9-f20b-4277-9629-15c3b95d12c4","Type":"ContainerStarted","Data":"d04171cd608c1757763503cb257c7c47790a6f20197fb32906b5987f3690b788"} Mar 20 09:52:02 crc kubenswrapper[4858]: I0320 09:52:02.586771 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29566672-gf59n" podStartSLOduration=1.76348471 podStartE2EDuration="2.586751872s" podCreationTimestamp="2026-03-20 09:52:00 +0000 UTC" firstStartedPulling="2026-03-20 09:52:01.02796895 +0000 UTC m=+3302.348387177" lastFinishedPulling="2026-03-20 09:52:01.851236102 +0000 UTC m=+3303.171654339" observedRunningTime="2026-03-20 09:52:02.582556057 +0000 UTC m=+3303.902974254" watchObservedRunningTime="2026-03-20 09:52:02.586751872 +0000 UTC m=+3303.907170079" Mar 20 09:52:03 crc kubenswrapper[4858]: I0320 09:52:03.573587 4858 generic.go:334] "Generic (PLEG): container finished" podID="5d3050c9-f20b-4277-9629-15c3b95d12c4" containerID="d04171cd608c1757763503cb257c7c47790a6f20197fb32906b5987f3690b788" exitCode=0 Mar 20 09:52:03 crc kubenswrapper[4858]: I0320 09:52:03.573635 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-gf59n" event={"ID":"5d3050c9-f20b-4277-9629-15c3b95d12c4","Type":"ContainerDied","Data":"d04171cd608c1757763503cb257c7c47790a6f20197fb32906b5987f3690b788"} Mar 20 09:52:04 crc kubenswrapper[4858]: I0320 09:52:04.961173 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-gf59n" Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.137502 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgddw\" (UniqueName: \"kubernetes.io/projected/5d3050c9-f20b-4277-9629-15c3b95d12c4-kube-api-access-kgddw\") pod \"5d3050c9-f20b-4277-9629-15c3b95d12c4\" (UID: \"5d3050c9-f20b-4277-9629-15c3b95d12c4\") " Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.156780 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d3050c9-f20b-4277-9629-15c3b95d12c4-kube-api-access-kgddw" (OuterVolumeSpecName: "kube-api-access-kgddw") pod "5d3050c9-f20b-4277-9629-15c3b95d12c4" (UID: "5d3050c9-f20b-4277-9629-15c3b95d12c4"). InnerVolumeSpecName "kube-api-access-kgddw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.239789 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgddw\" (UniqueName: \"kubernetes.io/projected/5d3050c9-f20b-4277-9629-15c3b95d12c4-kube-api-access-kgddw\") on node \"crc\" DevicePath \"\"" Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.591270 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566672-gf59n" event={"ID":"5d3050c9-f20b-4277-9629-15c3b95d12c4","Type":"ContainerDied","Data":"63c570fcf14c481e3ab1693496d298a87eecf1d6083b0fc3573a5c6f4a600b5b"} Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.591344 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63c570fcf14c481e3ab1693496d298a87eecf1d6083b0fc3573a5c6f4a600b5b" Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.591812 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566672-gf59n" Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.687687 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-69bjt"] Mar 20 09:52:05 crc kubenswrapper[4858]: I0320 09:52:05.698439 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566666-69bjt"] Mar 20 09:52:06 crc kubenswrapper[4858]: I0320 09:52:06.082394 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72eeae25-efc1-4285-8127-173c644400ab" path="/var/lib/kubelet/pods/72eeae25-efc1-4285-8127-173c644400ab/volumes" Mar 20 09:52:12 crc kubenswrapper[4858]: I0320 09:52:12.072076 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:52:12 crc kubenswrapper[4858]: E0320 09:52:12.072823 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:52:24 crc kubenswrapper[4858]: I0320 09:52:24.787025 4858 generic.go:334] "Generic (PLEG): container finished" podID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerID="cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a" exitCode=0 Mar 20 09:52:24 crc kubenswrapper[4858]: I0320 09:52:24.787191 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-m7hst/must-gather-lqx8h" event={"ID":"60bf6a80-0c17-41d8-a4af-46c4dc800572","Type":"ContainerDied","Data":"cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a"} Mar 20 09:52:24 crc kubenswrapper[4858]: I0320 09:52:24.788633 4858 scope.go:117] "RemoveContainer" containerID="cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a" Mar 20 09:52:25 crc kubenswrapper[4858]: I0320 09:52:25.070384 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:52:25 crc kubenswrapper[4858]: E0320 09:52:25.070841 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:52:25 crc kubenswrapper[4858]: I0320 09:52:25.333930 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m7hst_must-gather-lqx8h_60bf6a80-0c17-41d8-a4af-46c4dc800572/gather/0.log" Mar 20 09:52:31 crc kubenswrapper[4858]: I0320 09:52:31.951331 4858 scope.go:117] "RemoveContainer" containerID="e96491fec9fad04ed931f8100cf8790cc50bb31682e6425f655591b8a95f30f1" Mar 20 09:52:33 crc kubenswrapper[4858]: I0320 09:52:33.950282 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-m7hst/must-gather-lqx8h"] Mar 20 09:52:33 crc kubenswrapper[4858]: I0320 09:52:33.950959 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-m7hst/must-gather-lqx8h" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerName="copy" containerID="cri-o://3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1" gracePeriod=2 Mar 20 09:52:33 crc kubenswrapper[4858]: I0320 09:52:33.964633 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-m7hst/must-gather-lqx8h"] Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.417366 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m7hst_must-gather-lqx8h_60bf6a80-0c17-41d8-a4af-46c4dc800572/copy/0.log" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.418447 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.493121 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60bf6a80-0c17-41d8-a4af-46c4dc800572-must-gather-output\") pod \"60bf6a80-0c17-41d8-a4af-46c4dc800572\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.493248 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjbhw\" (UniqueName: \"kubernetes.io/projected/60bf6a80-0c17-41d8-a4af-46c4dc800572-kube-api-access-mjbhw\") pod \"60bf6a80-0c17-41d8-a4af-46c4dc800572\" (UID: \"60bf6a80-0c17-41d8-a4af-46c4dc800572\") " Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.498989 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60bf6a80-0c17-41d8-a4af-46c4dc800572-kube-api-access-mjbhw" (OuterVolumeSpecName: "kube-api-access-mjbhw") pod "60bf6a80-0c17-41d8-a4af-46c4dc800572" (UID: "60bf6a80-0c17-41d8-a4af-46c4dc800572"). InnerVolumeSpecName "kube-api-access-mjbhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.582898 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60bf6a80-0c17-41d8-a4af-46c4dc800572-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "60bf6a80-0c17-41d8-a4af-46c4dc800572" (UID: "60bf6a80-0c17-41d8-a4af-46c4dc800572"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.595877 4858 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/60bf6a80-0c17-41d8-a4af-46c4dc800572-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.595914 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjbhw\" (UniqueName: \"kubernetes.io/projected/60bf6a80-0c17-41d8-a4af-46c4dc800572-kube-api-access-mjbhw\") on node \"crc\" DevicePath \"\"" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.878591 4858 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-m7hst_must-gather-lqx8h_60bf6a80-0c17-41d8-a4af-46c4dc800572/copy/0.log" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.878980 4858 generic.go:334] "Generic (PLEG): container finished" podID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerID="3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1" exitCode=143 Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.879055 4858 scope.go:117] "RemoveContainer" containerID="3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.879217 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-m7hst/must-gather-lqx8h" Mar 20 09:52:34 crc kubenswrapper[4858]: I0320 09:52:34.924557 4858 scope.go:117] "RemoveContainer" containerID="cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a" Mar 20 09:52:35 crc kubenswrapper[4858]: I0320 09:52:35.010912 4858 scope.go:117] "RemoveContainer" containerID="3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1" Mar 20 09:52:35 crc kubenswrapper[4858]: E0320 09:52:35.011567 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1\": container with ID starting with 3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1 not found: ID does not exist" containerID="3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1" Mar 20 09:52:35 crc kubenswrapper[4858]: I0320 09:52:35.011595 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1"} err="failed to get container status \"3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1\": rpc error: code = NotFound desc = could not find container \"3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1\": container with ID starting with 3a8033f5d49cef59f2d971421238e4fa86e16351286077094428b7f847d749e1 not found: ID does not exist" Mar 20 09:52:35 crc kubenswrapper[4858]: I0320 09:52:35.011620 4858 scope.go:117] "RemoveContainer" containerID="cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a" Mar 20 09:52:35 crc kubenswrapper[4858]: E0320 09:52:35.012133 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a\": container with ID starting with cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a not found: ID does not exist" containerID="cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a" Mar 20 09:52:35 crc kubenswrapper[4858]: I0320 09:52:35.012153 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a"} err="failed to get container status \"cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a\": rpc error: code = NotFound desc = could not find container \"cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a\": container with ID starting with cc3ffb41ad8802b214f4185526d67b2208ecf9d4c40eb4edfe3874646c3ac82a not found: ID does not exist" Mar 20 09:52:36 crc kubenswrapper[4858]: I0320 09:52:36.082127 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" path="/var/lib/kubelet/pods/60bf6a80-0c17-41d8-a4af-46c4dc800572/volumes" Mar 20 09:52:38 crc kubenswrapper[4858]: I0320 09:52:38.070823 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:52:38 crc kubenswrapper[4858]: E0320 09:52:38.071710 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:52:52 crc kubenswrapper[4858]: I0320 09:52:52.070902 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:52:52 crc kubenswrapper[4858]: E0320 09:52:52.071759 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:53:06 crc kubenswrapper[4858]: I0320 09:53:06.070713 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:53:06 crc kubenswrapper[4858]: E0320 09:53:06.071899 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:53:19 crc kubenswrapper[4858]: I0320 09:53:19.070458 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:53:19 crc kubenswrapper[4858]: E0320 09:53:19.071150 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:53:30 crc kubenswrapper[4858]: I0320 09:53:30.078858 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:53:30 crc kubenswrapper[4858]: E0320 09:53:30.079839 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:53:45 crc kubenswrapper[4858]: I0320 09:53:45.072353 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:53:45 crc kubenswrapper[4858]: E0320 09:53:45.073110 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:53:58 crc kubenswrapper[4858]: I0320 09:53:58.070553 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:53:58 crc kubenswrapper[4858]: E0320 09:53:58.071811 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.154872 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566674-nvg2n"] Mar 20 09:54:00 crc kubenswrapper[4858]: E0320 09:54:00.155536 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerName="gather" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.155552 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerName="gather" Mar 20 09:54:00 crc kubenswrapper[4858]: E0320 09:54:00.155578 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerName="copy" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.155587 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerName="copy" Mar 20 09:54:00 crc kubenswrapper[4858]: E0320 09:54:00.155597 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d3050c9-f20b-4277-9629-15c3b95d12c4" containerName="oc" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.155604 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d3050c9-f20b-4277-9629-15c3b95d12c4" containerName="oc" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.155806 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerName="gather" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.155829 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d3050c9-f20b-4277-9629-15c3b95d12c4" containerName="oc" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.155838 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="60bf6a80-0c17-41d8-a4af-46c4dc800572" containerName="copy" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.156228 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566674-nvg2n" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.202087 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.202126 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.202257 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.215665 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t427\" (UniqueName: \"kubernetes.io/projected/063cd048-c953-4565-9691-168c3824847f-kube-api-access-4t427\") pod \"auto-csr-approver-29566674-nvg2n\" (UID: \"063cd048-c953-4565-9691-168c3824847f\") " pod="openshift-infra/auto-csr-approver-29566674-nvg2n" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.216997 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566674-nvg2n"] Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.317100 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t427\" (UniqueName: \"kubernetes.io/projected/063cd048-c953-4565-9691-168c3824847f-kube-api-access-4t427\") pod \"auto-csr-approver-29566674-nvg2n\" (UID: \"063cd048-c953-4565-9691-168c3824847f\") " pod="openshift-infra/auto-csr-approver-29566674-nvg2n" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.336174 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t427\" (UniqueName: \"kubernetes.io/projected/063cd048-c953-4565-9691-168c3824847f-kube-api-access-4t427\") pod \"auto-csr-approver-29566674-nvg2n\" (UID: \"063cd048-c953-4565-9691-168c3824847f\") " pod="openshift-infra/auto-csr-approver-29566674-nvg2n" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.514535 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566674-nvg2n" Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.967941 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566674-nvg2n"] Mar 20 09:54:00 crc kubenswrapper[4858]: I0320 09:54:00.970946 4858 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 20 09:54:01 crc kubenswrapper[4858]: I0320 09:54:01.583997 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566674-nvg2n" event={"ID":"063cd048-c953-4565-9691-168c3824847f","Type":"ContainerStarted","Data":"e49461eda2b73e94b15c1c7516786bb8ecf34c0190aa52b97efd30957ceb111a"} Mar 20 09:54:02 crc kubenswrapper[4858]: I0320 09:54:02.596819 4858 generic.go:334] "Generic (PLEG): container finished" podID="063cd048-c953-4565-9691-168c3824847f" containerID="f3ab82192a91fd56b2db836095ab5a3aeef14bd1bc3362863b3da6d456e8d442" exitCode=0 Mar 20 09:54:02 crc kubenswrapper[4858]: I0320 09:54:02.597053 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566674-nvg2n" event={"ID":"063cd048-c953-4565-9691-168c3824847f","Type":"ContainerDied","Data":"f3ab82192a91fd56b2db836095ab5a3aeef14bd1bc3362863b3da6d456e8d442"} Mar 20 09:54:03 crc kubenswrapper[4858]: I0320 09:54:03.978711 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566674-nvg2n" Mar 20 09:54:04 crc kubenswrapper[4858]: I0320 09:54:04.073977 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t427\" (UniqueName: \"kubernetes.io/projected/063cd048-c953-4565-9691-168c3824847f-kube-api-access-4t427\") pod \"063cd048-c953-4565-9691-168c3824847f\" (UID: \"063cd048-c953-4565-9691-168c3824847f\") " Mar 20 09:54:04 crc kubenswrapper[4858]: I0320 09:54:04.086615 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/063cd048-c953-4565-9691-168c3824847f-kube-api-access-4t427" (OuterVolumeSpecName: "kube-api-access-4t427") pod "063cd048-c953-4565-9691-168c3824847f" (UID: "063cd048-c953-4565-9691-168c3824847f"). InnerVolumeSpecName "kube-api-access-4t427". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:54:04 crc kubenswrapper[4858]: I0320 09:54:04.175705 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t427\" (UniqueName: \"kubernetes.io/projected/063cd048-c953-4565-9691-168c3824847f-kube-api-access-4t427\") on node \"crc\" DevicePath \"\"" Mar 20 09:54:04 crc kubenswrapper[4858]: I0320 09:54:04.616770 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566674-nvg2n" event={"ID":"063cd048-c953-4565-9691-168c3824847f","Type":"ContainerDied","Data":"e49461eda2b73e94b15c1c7516786bb8ecf34c0190aa52b97efd30957ceb111a"} Mar 20 09:54:04 crc kubenswrapper[4858]: I0320 09:54:04.617217 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e49461eda2b73e94b15c1c7516786bb8ecf34c0190aa52b97efd30957ceb111a" Mar 20 09:54:04 crc kubenswrapper[4858]: I0320 09:54:04.616789 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566674-nvg2n" Mar 20 09:54:05 crc kubenswrapper[4858]: I0320 09:54:05.051229 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566668-n2gcc"] Mar 20 09:54:05 crc kubenswrapper[4858]: I0320 09:54:05.061248 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566668-n2gcc"] Mar 20 09:54:06 crc kubenswrapper[4858]: I0320 09:54:06.081089 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cdceba8-aa1f-4931-930f-612d6ca0161d" path="/var/lib/kubelet/pods/1cdceba8-aa1f-4931-930f-612d6ca0161d/volumes" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.461420 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8tdtk"] Mar 20 09:54:12 crc kubenswrapper[4858]: E0320 09:54:12.462619 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063cd048-c953-4565-9691-168c3824847f" containerName="oc" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.462642 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="063cd048-c953-4565-9691-168c3824847f" containerName="oc" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.462894 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="063cd048-c953-4565-9691-168c3824847f" containerName="oc" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.464699 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.478905 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8tdtk"] Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.507579 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-catalog-content\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.507655 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-utilities\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.507691 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvl4t\" (UniqueName: \"kubernetes.io/projected/b375a437-05e7-466e-a1eb-341264fa5998-kube-api-access-vvl4t\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.608754 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-utilities\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.608826 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvl4t\" (UniqueName: \"kubernetes.io/projected/b375a437-05e7-466e-a1eb-341264fa5998-kube-api-access-vvl4t\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.608894 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-catalog-content\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.609355 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-utilities\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.609421 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-catalog-content\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.630455 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvl4t\" (UniqueName: \"kubernetes.io/projected/b375a437-05e7-466e-a1eb-341264fa5998-kube-api-access-vvl4t\") pod \"redhat-operators-8tdtk\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:12 crc kubenswrapper[4858]: I0320 09:54:12.782226 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:13 crc kubenswrapper[4858]: I0320 09:54:13.069998 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:54:13 crc kubenswrapper[4858]: E0320 09:54:13.070415 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:54:13 crc kubenswrapper[4858]: I0320 09:54:13.212916 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8tdtk"] Mar 20 09:54:13 crc kubenswrapper[4858]: I0320 09:54:13.687232 4858 generic.go:334] "Generic (PLEG): container finished" podID="b375a437-05e7-466e-a1eb-341264fa5998" containerID="e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370" exitCode=0 Mar 20 09:54:13 crc kubenswrapper[4858]: I0320 09:54:13.687282 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8tdtk" event={"ID":"b375a437-05e7-466e-a1eb-341264fa5998","Type":"ContainerDied","Data":"e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370"} Mar 20 09:54:13 crc kubenswrapper[4858]: I0320 09:54:13.687568 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8tdtk" event={"ID":"b375a437-05e7-466e-a1eb-341264fa5998","Type":"ContainerStarted","Data":"4c3ac860c1d7e7a053432e06228065f0c6c00c7fd061366babe5ebd676bde8b0"} Mar 20 09:54:14 crc kubenswrapper[4858]: I0320 09:54:14.697609 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8tdtk" event={"ID":"b375a437-05e7-466e-a1eb-341264fa5998","Type":"ContainerStarted","Data":"e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4"} Mar 20 09:54:15 crc kubenswrapper[4858]: I0320 09:54:15.709351 4858 generic.go:334] "Generic (PLEG): container finished" podID="b375a437-05e7-466e-a1eb-341264fa5998" containerID="e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4" exitCode=0 Mar 20 09:54:15 crc kubenswrapper[4858]: I0320 09:54:15.709598 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8tdtk" event={"ID":"b375a437-05e7-466e-a1eb-341264fa5998","Type":"ContainerDied","Data":"e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4"} Mar 20 09:54:16 crc kubenswrapper[4858]: I0320 09:54:16.722137 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8tdtk" event={"ID":"b375a437-05e7-466e-a1eb-341264fa5998","Type":"ContainerStarted","Data":"06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6"} Mar 20 09:54:16 crc kubenswrapper[4858]: I0320 09:54:16.753568 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8tdtk" podStartSLOduration=2.361457826 podStartE2EDuration="4.753547274s" podCreationTimestamp="2026-03-20 09:54:12 +0000 UTC" firstStartedPulling="2026-03-20 09:54:13.689123352 +0000 UTC m=+3435.009541549" lastFinishedPulling="2026-03-20 09:54:16.08121278 +0000 UTC m=+3437.401630997" observedRunningTime="2026-03-20 09:54:16.745804574 +0000 UTC m=+3438.066222771" watchObservedRunningTime="2026-03-20 09:54:16.753547274 +0000 UTC m=+3438.073965491" Mar 20 09:54:22 crc kubenswrapper[4858]: I0320 09:54:22.783167 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:22 crc kubenswrapper[4858]: I0320 09:54:22.783776 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:23 crc kubenswrapper[4858]: I0320 09:54:23.852729 4858 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8tdtk" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="registry-server" probeResult="failure" output=< Mar 20 09:54:23 crc kubenswrapper[4858]: timeout: failed to connect service ":50051" within 1s Mar 20 09:54:23 crc kubenswrapper[4858]: > Mar 20 09:54:27 crc kubenswrapper[4858]: I0320 09:54:27.070659 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:54:27 crc kubenswrapper[4858]: E0320 09:54:27.071350 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:54:32 crc kubenswrapper[4858]: I0320 09:54:32.060127 4858 scope.go:117] "RemoveContainer" containerID="cff5d1b3d6c4f1e18e142ef5b564e2ea4cab3e585dc82acaa59ff109c856cb2d" Mar 20 09:54:32 crc kubenswrapper[4858]: I0320 09:54:32.833751 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:32 crc kubenswrapper[4858]: I0320 09:54:32.903444 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:33 crc kubenswrapper[4858]: I0320 09:54:33.076896 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8tdtk"] Mar 20 09:54:33 crc kubenswrapper[4858]: I0320 09:54:33.875677 4858 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8tdtk" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="registry-server" containerID="cri-o://06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6" gracePeriod=2 Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.320951 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.449216 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-catalog-content\") pod \"b375a437-05e7-466e-a1eb-341264fa5998\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.449301 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvl4t\" (UniqueName: \"kubernetes.io/projected/b375a437-05e7-466e-a1eb-341264fa5998-kube-api-access-vvl4t\") pod \"b375a437-05e7-466e-a1eb-341264fa5998\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.449397 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-utilities\") pod \"b375a437-05e7-466e-a1eb-341264fa5998\" (UID: \"b375a437-05e7-466e-a1eb-341264fa5998\") " Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.450848 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-utilities" (OuterVolumeSpecName: "utilities") pod "b375a437-05e7-466e-a1eb-341264fa5998" (UID: "b375a437-05e7-466e-a1eb-341264fa5998"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.454733 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b375a437-05e7-466e-a1eb-341264fa5998-kube-api-access-vvl4t" (OuterVolumeSpecName: "kube-api-access-vvl4t") pod "b375a437-05e7-466e-a1eb-341264fa5998" (UID: "b375a437-05e7-466e-a1eb-341264fa5998"). InnerVolumeSpecName "kube-api-access-vvl4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.551036 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvl4t\" (UniqueName: \"kubernetes.io/projected/b375a437-05e7-466e-a1eb-341264fa5998-kube-api-access-vvl4t\") on node \"crc\" DevicePath \"\"" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.551082 4858 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-utilities\") on node \"crc\" DevicePath \"\"" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.576687 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b375a437-05e7-466e-a1eb-341264fa5998" (UID: "b375a437-05e7-466e-a1eb-341264fa5998"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.651849 4858 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b375a437-05e7-466e-a1eb-341264fa5998-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.888065 4858 generic.go:334] "Generic (PLEG): container finished" podID="b375a437-05e7-466e-a1eb-341264fa5998" containerID="06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6" exitCode=0 Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.888157 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8tdtk" event={"ID":"b375a437-05e7-466e-a1eb-341264fa5998","Type":"ContainerDied","Data":"06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6"} Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.888556 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8tdtk" event={"ID":"b375a437-05e7-466e-a1eb-341264fa5998","Type":"ContainerDied","Data":"4c3ac860c1d7e7a053432e06228065f0c6c00c7fd061366babe5ebd676bde8b0"} Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.888228 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8tdtk" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.888595 4858 scope.go:117] "RemoveContainer" containerID="06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.931260 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8tdtk"] Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.932843 4858 scope.go:117] "RemoveContainer" containerID="e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.939205 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8tdtk"] Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.957034 4858 scope.go:117] "RemoveContainer" containerID="e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.986292 4858 scope.go:117] "RemoveContainer" containerID="06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6" Mar 20 09:54:34 crc kubenswrapper[4858]: E0320 09:54:34.986841 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6\": container with ID starting with 06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6 not found: ID does not exist" containerID="06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.986888 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6"} err="failed to get container status \"06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6\": rpc error: code = NotFound desc = could not find container \"06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6\": container with ID starting with 06092d8e9f18e54e26f58257e4b5f0ac0c90eeb167d27e0683b2880846c8e6d6 not found: ID does not exist" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.986923 4858 scope.go:117] "RemoveContainer" containerID="e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4" Mar 20 09:54:34 crc kubenswrapper[4858]: E0320 09:54:34.987591 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4\": container with ID starting with e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4 not found: ID does not exist" containerID="e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.987691 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4"} err="failed to get container status \"e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4\": rpc error: code = NotFound desc = could not find container \"e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4\": container with ID starting with e81bcda91c1d8a1de51ee035d7fa4b500f8857a46579ebde919eeff9c139bbb4 not found: ID does not exist" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.987774 4858 scope.go:117] "RemoveContainer" containerID="e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370" Mar 20 09:54:34 crc kubenswrapper[4858]: E0320 09:54:34.988887 4858 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370\": container with ID starting with e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370 not found: ID does not exist" containerID="e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370" Mar 20 09:54:34 crc kubenswrapper[4858]: I0320 09:54:34.988956 4858 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370"} err="failed to get container status \"e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370\": rpc error: code = NotFound desc = could not find container \"e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370\": container with ID starting with e4004874cd23f1a325e51bc6febc8095c530a7cdb8b62d5f321fbe2ff7965370 not found: ID does not exist" Mar 20 09:54:36 crc kubenswrapper[4858]: I0320 09:54:36.086149 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b375a437-05e7-466e-a1eb-341264fa5998" path="/var/lib/kubelet/pods/b375a437-05e7-466e-a1eb-341264fa5998/volumes" Mar 20 09:54:41 crc kubenswrapper[4858]: I0320 09:54:41.070534 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:54:41 crc kubenswrapper[4858]: E0320 09:54:41.071158 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:54:54 crc kubenswrapper[4858]: I0320 09:54:54.070471 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:54:54 crc kubenswrapper[4858]: E0320 09:54:54.071582 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:55:09 crc kubenswrapper[4858]: I0320 09:55:09.070758 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:55:09 crc kubenswrapper[4858]: E0320 09:55:09.071635 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:55:23 crc kubenswrapper[4858]: I0320 09:55:23.070053 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:55:23 crc kubenswrapper[4858]: E0320 09:55:23.071157 4858 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w6t79_openshift-machine-config-operator(584bd2e0-0786-4137-9674-790c8fb680c5)\"" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" podUID="584bd2e0-0786-4137-9674-790c8fb680c5" Mar 20 09:55:38 crc kubenswrapper[4858]: I0320 09:55:38.070906 4858 scope.go:117] "RemoveContainer" containerID="393789da6d3dcae735717937b893e4d8ef76ccd6acc582a66682063bcc144711" Mar 20 09:55:38 crc kubenswrapper[4858]: I0320 09:55:38.443615 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w6t79" event={"ID":"584bd2e0-0786-4137-9674-790c8fb680c5","Type":"ContainerStarted","Data":"6b627bff178f61597ee6c6ee25f37615553c040dd22333f756a4cdd5f6540625"} Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.512567 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m2wm8"] Mar 20 09:55:55 crc kubenswrapper[4858]: E0320 09:55:55.513896 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="extract-utilities" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.513931 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="extract-utilities" Mar 20 09:55:55 crc kubenswrapper[4858]: E0320 09:55:55.513978 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="registry-server" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.513995 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="registry-server" Mar 20 09:55:55 crc kubenswrapper[4858]: E0320 09:55:55.514028 4858 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="extract-content" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.514045 4858 state_mem.go:107] "Deleted CPUSet assignment" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="extract-content" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.514435 4858 memory_manager.go:354] "RemoveStaleState removing state" podUID="b375a437-05e7-466e-a1eb-341264fa5998" containerName="registry-server" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.516718 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.535185 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2wm8"] Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.635390 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5916f65e-e6dd-40cf-b0ef-787d0e088692-catalog-content\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.635488 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9clxz\" (UniqueName: \"kubernetes.io/projected/5916f65e-e6dd-40cf-b0ef-787d0e088692-kube-api-access-9clxz\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.635574 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5916f65e-e6dd-40cf-b0ef-787d0e088692-utilities\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.736919 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5916f65e-e6dd-40cf-b0ef-787d0e088692-utilities\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.737012 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5916f65e-e6dd-40cf-b0ef-787d0e088692-catalog-content\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.737046 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9clxz\" (UniqueName: \"kubernetes.io/projected/5916f65e-e6dd-40cf-b0ef-787d0e088692-kube-api-access-9clxz\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.737600 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5916f65e-e6dd-40cf-b0ef-787d0e088692-utilities\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.737692 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5916f65e-e6dd-40cf-b0ef-787d0e088692-catalog-content\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.768718 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9clxz\" (UniqueName: \"kubernetes.io/projected/5916f65e-e6dd-40cf-b0ef-787d0e088692-kube-api-access-9clxz\") pod \"community-operators-m2wm8\" (UID: \"5916f65e-e6dd-40cf-b0ef-787d0e088692\") " pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:55 crc kubenswrapper[4858]: I0320 09:55:55.836700 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:55:56 crc kubenswrapper[4858]: I0320 09:55:56.128719 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2wm8"] Mar 20 09:55:56 crc kubenswrapper[4858]: I0320 09:55:56.601132 4858 generic.go:334] "Generic (PLEG): container finished" podID="5916f65e-e6dd-40cf-b0ef-787d0e088692" containerID="3566237f9abe90da85e72f1445c1fff4b28f31d9fb988bcb543fb0d342371f78" exitCode=0 Mar 20 09:55:56 crc kubenswrapper[4858]: I0320 09:55:56.601248 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2wm8" event={"ID":"5916f65e-e6dd-40cf-b0ef-787d0e088692","Type":"ContainerDied","Data":"3566237f9abe90da85e72f1445c1fff4b28f31d9fb988bcb543fb0d342371f78"} Mar 20 09:55:56 crc kubenswrapper[4858]: I0320 09:55:56.601511 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2wm8" event={"ID":"5916f65e-e6dd-40cf-b0ef-787d0e088692","Type":"ContainerStarted","Data":"5d0475be2ff25de6a8411f16d084e8704346e8e46b6ad50edc2d2b15eb1d0138"} Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.146845 4858 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29566676-dw9qn"] Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.148284 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566676-dw9qn" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.151291 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.151389 4858 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-9k6zg" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.152465 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566676-dw9qn"] Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.152187 4858 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.306602 4858 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8btwq\" (UniqueName: \"kubernetes.io/projected/0a416c58-2871-4c34-ad97-8673b0ca844f-kube-api-access-8btwq\") pod \"auto-csr-approver-29566676-dw9qn\" (UID: \"0a416c58-2871-4c34-ad97-8673b0ca844f\") " pod="openshift-infra/auto-csr-approver-29566676-dw9qn" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.407864 4858 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8btwq\" (UniqueName: \"kubernetes.io/projected/0a416c58-2871-4c34-ad97-8673b0ca844f-kube-api-access-8btwq\") pod \"auto-csr-approver-29566676-dw9qn\" (UID: \"0a416c58-2871-4c34-ad97-8673b0ca844f\") " pod="openshift-infra/auto-csr-approver-29566676-dw9qn" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.426089 4858 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8btwq\" (UniqueName: \"kubernetes.io/projected/0a416c58-2871-4c34-ad97-8673b0ca844f-kube-api-access-8btwq\") pod \"auto-csr-approver-29566676-dw9qn\" (UID: \"0a416c58-2871-4c34-ad97-8673b0ca844f\") " pod="openshift-infra/auto-csr-approver-29566676-dw9qn" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.476966 4858 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566676-dw9qn" Mar 20 09:56:00 crc kubenswrapper[4858]: I0320 09:56:00.983896 4858 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29566676-dw9qn"] Mar 20 09:56:00 crc kubenswrapper[4858]: W0320 09:56:00.987889 4858 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a416c58_2871_4c34_ad97_8673b0ca844f.slice/crio-d4aef2cda62f9ac2b45de0ffac83fdb08169506bdcc1505e6e74336f11b61397 WatchSource:0}: Error finding container d4aef2cda62f9ac2b45de0ffac83fdb08169506bdcc1505e6e74336f11b61397: Status 404 returned error can't find the container with id d4aef2cda62f9ac2b45de0ffac83fdb08169506bdcc1505e6e74336f11b61397 Mar 20 09:56:01 crc kubenswrapper[4858]: I0320 09:56:01.636553 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566676-dw9qn" event={"ID":"0a416c58-2871-4c34-ad97-8673b0ca844f","Type":"ContainerStarted","Data":"d4aef2cda62f9ac2b45de0ffac83fdb08169506bdcc1505e6e74336f11b61397"} Mar 20 09:56:01 crc kubenswrapper[4858]: I0320 09:56:01.638699 4858 generic.go:334] "Generic (PLEG): container finished" podID="5916f65e-e6dd-40cf-b0ef-787d0e088692" containerID="e81c4554041941c92710e80df283837c84f8d569d12989185b7d01d1e9b16d61" exitCode=0 Mar 20 09:56:01 crc kubenswrapper[4858]: I0320 09:56:01.638732 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2wm8" event={"ID":"5916f65e-e6dd-40cf-b0ef-787d0e088692","Type":"ContainerDied","Data":"e81c4554041941c92710e80df283837c84f8d569d12989185b7d01d1e9b16d61"} Mar 20 09:56:02 crc kubenswrapper[4858]: I0320 09:56:02.646387 4858 generic.go:334] "Generic (PLEG): container finished" podID="0a416c58-2871-4c34-ad97-8673b0ca844f" containerID="bd6ab2fd4b7df6c87102c57f26939c51c4a9d5d3d350449bb9d28b28b3559006" exitCode=0 Mar 20 09:56:02 crc kubenswrapper[4858]: I0320 09:56:02.646445 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566676-dw9qn" event={"ID":"0a416c58-2871-4c34-ad97-8673b0ca844f","Type":"ContainerDied","Data":"bd6ab2fd4b7df6c87102c57f26939c51c4a9d5d3d350449bb9d28b28b3559006"} Mar 20 09:56:02 crc kubenswrapper[4858]: I0320 09:56:02.650517 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2wm8" event={"ID":"5916f65e-e6dd-40cf-b0ef-787d0e088692","Type":"ContainerStarted","Data":"7e96b0d209ed6ca3d6aba380c3b991167022649187bd653c3c5198412da0db39"} Mar 20 09:56:02 crc kubenswrapper[4858]: I0320 09:56:02.684565 4858 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m2wm8" podStartSLOduration=2.224405086 podStartE2EDuration="7.684545993s" podCreationTimestamp="2026-03-20 09:55:55 +0000 UTC" firstStartedPulling="2026-03-20 09:55:56.60504531 +0000 UTC m=+3537.925463547" lastFinishedPulling="2026-03-20 09:56:02.065186257 +0000 UTC m=+3543.385604454" observedRunningTime="2026-03-20 09:56:02.676556617 +0000 UTC m=+3543.996974814" watchObservedRunningTime="2026-03-20 09:56:02.684545993 +0000 UTC m=+3544.004964200" Mar 20 09:56:03 crc kubenswrapper[4858]: I0320 09:56:03.957873 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566676-dw9qn" Mar 20 09:56:04 crc kubenswrapper[4858]: I0320 09:56:04.062680 4858 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8btwq\" (UniqueName: \"kubernetes.io/projected/0a416c58-2871-4c34-ad97-8673b0ca844f-kube-api-access-8btwq\") pod \"0a416c58-2871-4c34-ad97-8673b0ca844f\" (UID: \"0a416c58-2871-4c34-ad97-8673b0ca844f\") " Mar 20 09:56:04 crc kubenswrapper[4858]: I0320 09:56:04.067834 4858 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a416c58-2871-4c34-ad97-8673b0ca844f-kube-api-access-8btwq" (OuterVolumeSpecName: "kube-api-access-8btwq") pod "0a416c58-2871-4c34-ad97-8673b0ca844f" (UID: "0a416c58-2871-4c34-ad97-8673b0ca844f"). InnerVolumeSpecName "kube-api-access-8btwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 20 09:56:04 crc kubenswrapper[4858]: I0320 09:56:04.164482 4858 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8btwq\" (UniqueName: \"kubernetes.io/projected/0a416c58-2871-4c34-ad97-8673b0ca844f-kube-api-access-8btwq\") on node \"crc\" DevicePath \"\"" Mar 20 09:56:04 crc kubenswrapper[4858]: I0320 09:56:04.667288 4858 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29566676-dw9qn" event={"ID":"0a416c58-2871-4c34-ad97-8673b0ca844f","Type":"ContainerDied","Data":"d4aef2cda62f9ac2b45de0ffac83fdb08169506bdcc1505e6e74336f11b61397"} Mar 20 09:56:04 crc kubenswrapper[4858]: I0320 09:56:04.667681 4858 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4aef2cda62f9ac2b45de0ffac83fdb08169506bdcc1505e6e74336f11b61397" Mar 20 09:56:04 crc kubenswrapper[4858]: I0320 09:56:04.667399 4858 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29566676-dw9qn" Mar 20 09:56:05 crc kubenswrapper[4858]: I0320 09:56:05.045743 4858 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29566670-6qb6v"] Mar 20 09:56:05 crc kubenswrapper[4858]: I0320 09:56:05.058558 4858 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29566670-6qb6v"] Mar 20 09:56:05 crc kubenswrapper[4858]: I0320 09:56:05.838125 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:56:05 crc kubenswrapper[4858]: I0320 09:56:05.838255 4858 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:56:05 crc kubenswrapper[4858]: I0320 09:56:05.898028 4858 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m2wm8" Mar 20 09:56:06 crc kubenswrapper[4858]: I0320 09:56:06.102213 4858 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6bb9eff-c7c8-4754-a74a-80e52fd3f103" path="/var/lib/kubelet/pods/e6bb9eff-c7c8-4754-a74a-80e52fd3f103/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515157214701024450 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015157214702017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015157205401016505 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015157205401015455 5ustar corecore